Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
207,876 | 15,855,835,399 | IssuesEvent | 2021-04-08 00:51:45 | backend-br/vagas | https://api.github.com/repos/backend-br/vagas | closed | [Remoto] Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz | AWS CI CLT Especialista GraphQL NoSQL Python Remoto Rest Stale Testes automatizados | ## Descrição da vaga:
<p>Nosso propósito aqui no Méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa Grande Família.</p>
<p>Estamos entre as melhores empresas para se trabalhar no ranking empresas de Tecnologia e no Ranking Minas Gerais, segundo o Great Place to Work. Isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores, promovendo o bem estar de todos que se relacionam conosco, com foco em objetivos desafiadores. Afinal, queremos alçar vôos muito mais altos.</p>
<p>Temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família. Especificamente para esta vaga, estamos em busca de pessoas de nível Sênior/Especialista em tecnologias Back-end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis. Temos produtos acessados por milhões de pessoas e muitos projetos para torná-los ainda mais interessantes para nossos usuários e parceiros.</p>
<p><strong>😏Já pensou em trabalhar remotamente, de qualquer lugar do Brasil ou de um dos nossos escritórios em Belo Horizonte ou Manaus, e ainda impactar positivamente na vida de milhões de usuários? 🙌</strong></p>
<p></p>
<p><span style="font-size: 18px;"><strong>RESPONSABILIDADES E ATRIBUIÇÕES</strong></span></p>
<ul>
<li>Colaborar com nossos times de engenharia, produto e negócio na construção dos nossos produtos;</li>
<li>Atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend;</li>
<li>Projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados;</li>
<li>Projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente;</li>
<li>Arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção;</li>
<li>Atuar como mentor e disseminar conhecimento para outros membros do time.</li>
</ul>
<p></p>
<p><strong>MAIS IMPORTANTE AINDA:</strong></p>
<p>👩💻Todas as nossas vagas estão disponíveis para pessoas com deficiência!</p>
## Méliuz:
<p><span style="color: rgb(30,32,34);background-color: rgb(255,255,255);font-size: 16px;font-family: Poppins, Helvetica, Arial, sans-serif;">O Méliuz é uma empresa que desbloqueia relações inteligentes de consumo. Estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras, de maneira que todos saiam ganhando. Queremos que nossos parceiros continuem conosco para toda a vida e, para isso, estamos sempre levando inovações até eles. Tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais.</span> </p>
</p>
## Habilidades:
- AWS
- Node.js
- CI/CD
- NoSQL
## Local:
100% Remoto
## Requisitos:
- Conhecimentos avançados em linguagens e tecnologias como Python e Node.js;
- Conhecimentos avançados de algoritmos e estruturas de dados;
- Experiência trabalhando com banco de dados relacionais e NoSQL;
- Experiência sólida desenvolvendo APIs com REST, gRPC ou GraphQL;
- Experiência utilizando cache (elasticache) e serviços de fila(SQS/SNS);
- Arquitetura orientada a eventos e microsserviços;
- Experiência com testes automatizados e ambientes de CI/CD;
- Experiência implementando arquiteturas escaláveis em ambiente cloud, principalmente AWS.
## Benefícios:
- 🚈 Vale transporte;
- 🥗 Vale refeição ou alimentação;
- 🤩Plano de saúde;
- 😁 Plano odontológico;
- ⏰ Horário flexível;
- 💰 Premiação por atingimento de metas coletivas;
- 🏊♀ Convênio com Sesc;
- 📚 Biblioteca interna;
- 🤑 Cashback online em dobro;
- 🤰 Licença Maternidade/Paternidade estendida;
- 💻 Ajuda de custo para o teletrabalho;
- 👶 Auxílio Creche;
- ✝ Auxílio luto.
## Como se candidatar:
Candidate-se em: [Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz](https://coodesh.com/vagas/pessoa-desenvolvedora-backend-seniorespecialista-vaga-remota-002652?origin=github&modal=open)
## Tempo médio de feedbacks:
Costumamos enviar feedbacks em até 48 horas após cada etapa do processo. E-mail para contato em caso de não haver resposta: [tamara.carvalho@meliuz.com.br](mailto:tamara.carvalho@meliuz.com.br)
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Nível
Sênior | 1.0 | [Remoto] Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz - ## Descrição da vaga:
<p>Nosso propósito aqui no Méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa Grande Família.</p>
<p>Estamos entre as melhores empresas para se trabalhar no ranking empresas de Tecnologia e no Ranking Minas Gerais, segundo o Great Place to Work. Isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores, promovendo o bem estar de todos que se relacionam conosco, com foco em objetivos desafiadores. Afinal, queremos alçar vôos muito mais altos.</p>
<p>Temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família. Especificamente para esta vaga, estamos em busca de pessoas de nível Sênior/Especialista em tecnologias Back-end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis. Temos produtos acessados por milhões de pessoas e muitos projetos para torná-los ainda mais interessantes para nossos usuários e parceiros.</p>
<p><strong>😏Já pensou em trabalhar remotamente, de qualquer lugar do Brasil ou de um dos nossos escritórios em Belo Horizonte ou Manaus, e ainda impactar positivamente na vida de milhões de usuários? 🙌</strong></p>
<p></p>
<p><span style="font-size: 18px;"><strong>RESPONSABILIDADES E ATRIBUIÇÕES</strong></span></p>
<ul>
<li>Colaborar com nossos times de engenharia, produto e negócio na construção dos nossos produtos;</li>
<li>Atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend;</li>
<li>Projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados;</li>
<li>Projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente;</li>
<li>Arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção;</li>
<li>Atuar como mentor e disseminar conhecimento para outros membros do time.</li>
</ul>
<p></p>
<p><strong>MAIS IMPORTANTE AINDA:</strong></p>
<p>👩💻Todas as nossas vagas estão disponíveis para pessoas com deficiência!</p>
## Méliuz:
<p><span style="color: rgb(30,32,34);background-color: rgb(255,255,255);font-size: 16px;font-family: Poppins, Helvetica, Arial, sans-serif;">O Méliuz é uma empresa que desbloqueia relações inteligentes de consumo. Estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras, de maneira que todos saiam ganhando. Queremos que nossos parceiros continuem conosco para toda a vida e, para isso, estamos sempre levando inovações até eles. Tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais.</span> </p>
</p>
## Habilidades:
- AWS
- Node.js
- CI/CD
- NoSQL
## Local:
100% Remoto
## Requisitos:
- Conhecimentos avançados em linguagens e tecnologias como Python e Node.js;
- Conhecimentos avançados de algoritmos e estruturas de dados;
- Experiência trabalhando com banco de dados relacionais e NoSQL;
- Experiência sólida desenvolvendo APIs com REST, gRPC ou GraphQL;
- Experiência utilizando cache (elasticache) e serviços de fila(SQS/SNS);
- Arquitetura orientada a eventos e microsserviços;
- Experiência com testes automatizados e ambientes de CI/CD;
- Experiência implementando arquiteturas escaláveis em ambiente cloud, principalmente AWS.
## Benefícios:
- 🚈 Vale transporte;
- 🥗 Vale refeição ou alimentação;
- 🤩Plano de saúde;
- 😁 Plano odontológico;
- ⏰ Horário flexível;
- 💰 Premiação por atingimento de metas coletivas;
- 🏊♀ Convênio com Sesc;
- 📚 Biblioteca interna;
- 🤑 Cashback online em dobro;
- 🤰 Licença Maternidade/Paternidade estendida;
- 💻 Ajuda de custo para o teletrabalho;
- 👶 Auxílio Creche;
- ✝ Auxílio luto.
## Como se candidatar:
Candidate-se em: [Pessoa Desenvolvedora Back-end Sênior/Especialista (Vaga Remota) na Méliuz](https://coodesh.com/vagas/pessoa-desenvolvedora-backend-seniorespecialista-vaga-remota-002652?origin=github&modal=open)
## Tempo médio de feedbacks:
Costumamos enviar feedbacks em até 48 horas após cada etapa do processo. E-mail para contato em caso de não haver resposta: [tamara.carvalho@meliuz.com.br](mailto:tamara.carvalho@meliuz.com.br)
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Nível
Sênior | test | pessoa desenvolvedora back end sênior especialista vaga remota na méliuz descrição da vaga nosso propósito aqui no méliuz é desbloquear relações inteligentes de consumo e para fazermos isso todos os dias contamos com a nossa grande família estamos entre as melhores empresas para se trabalhar no ranking empresas de tecnologia e no ranking minas gerais segundo o great place to work isso significa que temos o compromisso de conduzir processos alinhados com os nossos valores promovendo o bem estar de todos que se relacionam conosco com foco em objetivos desafiadores afinal queremos alçar vôos muito mais altos temos grandes sonhos e desafios e sabemos que para chegarmos cada vez mais longe precisamos aumentar nossa grande família especificamente para esta vaga estamos em busca de pessoas de nível sênior especialista em tecnologias back end que nos ajudarão a continuar aprimorando nossa arquitetura voltada a microsserviços e a construir aplicações cada vez mais escaláveis temos produtos acessados por milhões de pessoas e muitos projetos para torná los ainda mais interessantes para nossos usuários e parceiros 😏já pensou em trabalhar remotamente de qualquer lugar do brasil ou de um dos nossos escritórios em belo horizonte ou manaus e ainda impactar positivamente na vida de milhões de usuários 🙌 responsabilidades e atribuições colaborar com nossos times de engenharia produto e negócio na construção dos nossos produtos atuar com protagonismo no desenvolvimento e evolução das nossas tecnologias backend projetar soluções e arquiteturas escaláveis para atender grandes volumes de acessos e tráfego de dados projetar e implementar a persistência de dados das aplicações e garantir que o acesso seja seguro e eficiente arquitetar soluções para problemas complexos e garantir que o código seja eficiente e de fácil manutenção atuar como mentor e disseminar conhecimento para outros membros do time mais importante ainda 👩💻todas as nossas vagas estão disponíveis para pessoas com deficiência méliuz o méliuz é uma empresa que desbloqueia relações inteligentes de consumo estamos em uma busca constante de melhorar a forma que nossos usuários fazem suas compras em nossas lojas parceiras de maneira que todos saiam ganhando queremos que nossos parceiros continuem conosco para toda a vida e para isso estamos sempre levando inovações até eles tudo isso aliado ao nosso espírito empreendedor que nos faz evoluir cada dia mais nbsp habilidades aws node js ci cd nosql local remoto requisitos conhecimentos avançados em linguagens e tecnologias como python e node js conhecimentos avançados de algoritmos e estruturas de dados experiência trabalhando com banco de dados relacionais e nosql experiência sólida desenvolvendo apis com rest grpc ou graphql experiência utilizando cache elasticache e serviços de fila sqs sns arquitetura orientada a eventos e microsserviços experiência com testes automatizados e ambientes de ci cd experiência implementando arquiteturas escaláveis em ambiente cloud principalmente aws benefícios 🚈 vale transporte 🥗 vale refeição ou alimentação 🤩plano de saúde 😁 plano odontológico ⏰ horário flexível 💰 premiação por atingimento de metas coletivas 🏊♀ convênio com sesc 📚 biblioteca interna 🤑 cashback online em dobro 🤰 licença maternidade paternidade estendida 💻 ajuda de custo para o teletrabalho 👶 auxílio creche ✝ auxílio luto como se candidatar candidate se em tempo médio de feedbacks costumamos enviar feedbacks em até horas após cada etapa do processo e mail para contato em caso de não haver resposta mailto tamara carvalho meliuz com br labels alocação remoto regime clt nível sênior | 1 |
158,691 | 12,422,628,872 | IssuesEvent | 2020-05-23 23:29:11 | sqlalchemy/sqlalchemy | https://api.github.com/repos/sqlalchemy/sqlalchemy | closed | Update profiles to python 3.8 | tests | The ci now uses python 3.8 but the profiles are still for python 3.7. | 1.0 | Update profiles to python 3.8 - The ci now uses python 3.8 but the profiles are still for python 3.7. | test | update profiles to python the ci now uses python but the profiles are still for python | 1 |
419,862 | 12,229,305,440 | IssuesEvent | 2020-05-03 23:31:46 | roed314/seminars | https://api.github.com/repos/roed314/seminars | opened | Declutter view talk page | low priority user interface | On each "View talk" page:
1) The "Livestream access" box does not deserve to have its own line. It should be moved to the right of the text "Mon May 4, 07:30-08:30 (starts in 12 hours)".
2) Maybe we could remove the word "Abstract" since in essentially all cases, it is obvious that the abstract is the abstract.
3) I think we could simplify
Talk content
* paper
* slides
* video
to
(paper | slides | video)
which simplifies further to
(slides | video)
(slides)
or nothing at all, depending on how many items are present.
Or if we really want to keep the words "Talk content", at least put everything on one line:
Talk content: (paper | slides | video)
| 1.0 | Declutter view talk page - On each "View talk" page:
1) The "Livestream access" box does not deserve to have its own line. It should be moved to the right of the text "Mon May 4, 07:30-08:30 (starts in 12 hours)".
2) Maybe we could remove the word "Abstract" since in essentially all cases, it is obvious that the abstract is the abstract.
3) I think we could simplify
Talk content
* paper
* slides
* video
to
(paper | slides | video)
which simplifies further to
(slides | video)
(slides)
or nothing at all, depending on how many items are present.
Or if we really want to keep the words "Talk content", at least put everything on one line:
Talk content: (paper | slides | video)
| non_test | declutter view talk page on each view talk page the livestream access box does not deserve to have its own line it should be moved to the right of the text mon may starts in hours maybe we could remove the word abstract since in essentially all cases it is obvious that the abstract is the abstract i think we could simplify talk content paper slides video to paper slides video which simplifies further to slides video slides or nothing at all depending on how many items are present or if we really want to keep the words talk content at least put everything on one line talk content paper slides video | 0 |
60,000 | 8,393,874,120 | IssuesEvent | 2018-10-09 21:57:10 | mfem/mfem | https://api.github.com/repos/mfem/mfem | closed | Nedelec bases on prisms and tetrahedra | documentation enhancement fem | Hello,
I need to have a discussion about high order Nedelec bases on prisms and tetrahedra. Mainly so that I'm sure I have a reasonable understanding of the requirements before I make another attempt at implementing ND and RT bases for prisms.
The ND basis functions associated with edges have a clearly defined orientation which is directed along the edge. Interior basis functions can be defined however we choose. Basis functions on quadrilateral faces can also borrow their orientation from the bounding edges. However, the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges.
The tetrahedral implementation in MFEM reorders the vertices of each tetrahedron so that the two lowest vertex indices are first. The ND basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face. This effectively determines the basis function orientations for every face in a tetrahedral mesh. Moreover, face 0 of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions, and likewise for faces 1, 2, and 3.
I don't believe there's a way to define prisms so that they posses this same convenient behavior. If we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism. In effect this produces three types of prisms in a general mesh.
This is unlike what we see for tetrahedra or hexahedra in MFEM. At the very least I expect I'll need to introduce more flexibility into how we handle DOF permutations when working with prisms. What I don't know is just how well I can contain these changes. Any thoughts on potential solutions or pitfalls would be appreciated.
Thanks!
Mark
| 1.0 | Nedelec bases on prisms and tetrahedra - Hello,
I need to have a discussion about high order Nedelec bases on prisms and tetrahedra. Mainly so that I'm sure I have a reasonable understanding of the requirements before I make another attempt at implementing ND and RT bases for prisms.
The ND basis functions associated with edges have a clearly defined orientation which is directed along the edge. Interior basis functions can be defined however we choose. Basis functions on quadrilateral faces can also borrow their orientation from the bounding edges. However, the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges.
The tetrahedral implementation in MFEM reorders the vertices of each tetrahedron so that the two lowest vertex indices are first. The ND basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face. This effectively determines the basis function orientations for every face in a tetrahedral mesh. Moreover, face 0 of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions, and likewise for faces 1, 2, and 3.
I don't believe there's a way to define prisms so that they posses this same convenient behavior. If we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism. In effect this produces three types of prisms in a general mesh.
This is unlike what we see for tetrahedra or hexahedra in MFEM. At the very least I expect I'll need to introduce more flexibility into how we handle DOF permutations when working with prisms. What I don't know is just how well I can contain these changes. Any thoughts on potential solutions or pitfalls would be appreciated.
Thanks!
Mark
| non_test | nedelec bases on prisms and tetrahedra hello i need to have a discussion about high order nedelec bases on prisms and tetrahedra mainly so that i m sure i have a reasonable understanding of the requirements before i make another attempt at implementing nd and rt bases for prisms the nd basis functions associated with edges have a clearly defined orientation which is directed along the edge interior basis functions can be defined however we choose basis functions on quadrilateral faces can also borrow their orientation from the bounding edges however the basis functions associated with triangular faces must align with two of the three bounding edges and its important that neighboring elements choose the same pair of edges the tetrahedral implementation in mfem reorders the vertices of each tetrahedron so that the two lowest vertex indices are first the nd basis functions on triangular faces are then defined so that they align with the two edges which meet at the vertex with the lowest index on each face this effectively determines the basis function orientations for every face in a tetrahedral mesh moreover face of every tetrahedron in the mesh will choose the same pair of local edges to define its basis functions and likewise for faces and i don t believe there s a way to define prisms so that they posses this same convenient behavior if we choose the basis functions on the base of the prism in a certain way there are still three possible choices for the orientation of the basis functions on the top of the prism in effect this produces three types of prisms in a general mesh this is unlike what we see for tetrahedra or hexahedra in mfem at the very least i expect i ll need to introduce more flexibility into how we handle dof permutations when working with prisms what i don t know is just how well i can contain these changes any thoughts on potential solutions or pitfalls would be appreciated thanks mark | 0 |
149,441 | 13,281,333,405 | IssuesEvent | 2020-08-23 16:50:54 | genedan/TmVal | https://api.github.com/repos/genedan/TmVal | closed | Rework simple interest examples in docs | documentation | Since SimpleAmt has been deprecated, we need new examples | 1.0 | Rework simple interest examples in docs - Since SimpleAmt has been deprecated, we need new examples | non_test | rework simple interest examples in docs since simpleamt has been deprecated we need new examples | 0 |
216,219 | 16,748,249,928 | IssuesEvent | 2021-06-11 18:34:10 | firebase/firebase-cpp-sdk | https://api.github.com/repos/firebase/firebase-cpp-sdk | closed | Nightly Integration Testing Report | nightly-testing tests: failed | ### ❌ Integration test FAILED
Requested by @sunmou99 on commit f7031c96a736f4bf4a56c9a164f261f9ec0d80e4
Last updated: Fri Jun 11 02:42 PDT 2021
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/927716699)**
| Failures | Configs |
|----------|---------|
| auth | <details><summary>(5 items)</summary>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/>[TEST] [FAILURE] [Android] [macos, windows] [emulator_min]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Android] [ubuntu] [emulator_target]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [MacOS] [All ssl_lib]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Windows] [openssl]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details></details> |
| database | [TEST] [ERROR] [iOS] [macos] [All Simulators]<br/> |
| dynamic_links | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest]<br/> |
| firestore | <details><summary>(4 items)</summary>[TEST] [ERROR] [Android] [All os] [emulator_min, android_min]<br/>[TEST] [ERROR] [Android] [macos] [emulator_target]<br/>[TEST] [ERROR] [Android] [ubuntu] [android_target]<br/>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/></details> |
| functions | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(7 failed tests)</summary> TestErrorHandling<br/> TestFunction<br/> TestFunctionWithAuthToken<br/> TestFunctionWithData<br/> TestFunctionWithNull<br/> TestFunctionWithScalar<br/> TestSignIn</details> |
| installations | [TEST] [ERROR] [iOS] [macos] [simulator_target]<br/>[TEST] [FAILURE] [Android] [windows] [android_min]<details><summary>(1 failed tests)</summary> TestGettingTokenTwiceMatches</details>[TEST] [FAILURE] [iOS] [macos] [simulator_latest, simulator_min]<details><summary>(8 failed tests)</summary> TestCanGetId<br/> TestCanGetIdAndTokenTogether<br/> TestCanGetToken<br/> TestDeleteGivesNewIdNextTime<br/> TestDeleteGivesNewTokenNextTime<br/> TestGetTokenForceRefresh<br/> TestGettingIdTwiceMatches<br/> TestGettingTokenTwiceMatches</details> |
| messaging | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest, android_target, android_min]<br/> |
| remote_config | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(2 failed tests)</summary> TestFetchV2<br/> TestGetAll</details> |
| storage | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(12 failed tests)</summary> TestCreateWorkingFolder<br/> TestDeleteFile<br/> TestDownloadUrl<br/> TestInvalidatingReferencesWhenDeletingApp<br/> TestInvalidatingReferencesWhenDeletingStorage<br/> TestLargeFileCancelUpload<br/> TestLargeFilePauseResumeAndDownloadCancel<br/> TestPutFileAndGetFile<br/> TestSignIn<br/> TestStorageUrl<br/> TestWriteAndReadByteBuffer<br/> TestWriteAndReadFileWithCustomMetadata</details> |
<hidden value="integration-test-status-comment"></hidden> | 2.0 | Nightly Integration Testing Report - ### ❌ Integration test FAILED
Requested by @sunmou99 on commit f7031c96a736f4bf4a56c9a164f261f9ec0d80e4
Last updated: Fri Jun 11 02:42 PDT 2021
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/927716699)**
| Failures | Configs |
|----------|---------|
| auth | <details><summary>(5 items)</summary>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/>[TEST] [FAILURE] [Android] [macos, windows] [emulator_min]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Android] [ubuntu] [emulator_target]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [MacOS] [All ssl_lib]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details>[TEST] [FAILURE] [Windows] [openssl]<details><summary>(1 failed tests)</summary> TestSendPasswordResetEmail</details></details> |
| database | [TEST] [ERROR] [iOS] [macos] [All Simulators]<br/> |
| dynamic_links | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest]<br/> |
| firestore | <details><summary>(4 items)</summary>[TEST] [ERROR] [Android] [All os] [emulator_min, android_min]<br/>[TEST] [ERROR] [Android] [macos] [emulator_target]<br/>[TEST] [ERROR] [Android] [ubuntu] [android_target]<br/>[TEST] [ERROR] [iOS] [macos] [All Simulators]<br/></details> |
| functions | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(7 failed tests)</summary> TestErrorHandling<br/> TestFunction<br/> TestFunctionWithAuthToken<br/> TestFunctionWithData<br/> TestFunctionWithNull<br/> TestFunctionWithScalar<br/> TestSignIn</details> |
| installations | [TEST] [ERROR] [iOS] [macos] [simulator_target]<br/>[TEST] [FAILURE] [Android] [windows] [android_min]<details><summary>(1 failed tests)</summary> TestGettingTokenTwiceMatches</details>[TEST] [FAILURE] [iOS] [macos] [simulator_latest, simulator_min]<details><summary>(8 failed tests)</summary> TestCanGetId<br/> TestCanGetIdAndTokenTogether<br/> TestCanGetToken<br/> TestDeleteGivesNewIdNextTime<br/> TestDeleteGivesNewTokenNextTime<br/> TestGetTokenForceRefresh<br/> TestGettingIdTwiceMatches<br/> TestGettingTokenTwiceMatches</details> |
| messaging | [TEST] [ERROR] [Android] [All os] [emulator_min, emulator_latest, android_target, android_min]<br/> |
| remote_config | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(2 failed tests)</summary> TestFetchV2<br/> TestGetAll</details> |
| storage | [TEST] [FAILURE] [iOS] [macos] [All Simulators]<details><summary>(12 failed tests)</summary> TestCreateWorkingFolder<br/> TestDeleteFile<br/> TestDownloadUrl<br/> TestInvalidatingReferencesWhenDeletingApp<br/> TestInvalidatingReferencesWhenDeletingStorage<br/> TestLargeFileCancelUpload<br/> TestLargeFilePauseResumeAndDownloadCancel<br/> TestPutFileAndGetFile<br/> TestSignIn<br/> TestStorageUrl<br/> TestWriteAndReadByteBuffer<br/> TestWriteAndReadFileWithCustomMetadata</details> |
<hidden value="integration-test-status-comment"></hidden> | test | nightly integration testing report ❌ nbsp integration test failed requested by on commit last updated fri jun pdt failures configs auth items failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail failed tests nbsp nbsp testsendpasswordresetemail database dynamic links firestore items functions failed tests nbsp nbsp testerrorhandling nbsp nbsp testfunction nbsp nbsp testfunctionwithauthtoken nbsp nbsp testfunctionwithdata nbsp nbsp testfunctionwithnull nbsp nbsp testfunctionwithscalar nbsp nbsp testsignin installations failed tests nbsp nbsp testgettingtokentwicematches failed tests nbsp nbsp testcangetid nbsp nbsp testcangetidandtokentogether nbsp nbsp testcangettoken nbsp nbsp testdeletegivesnewidnexttime nbsp nbsp testdeletegivesnewtokennexttime nbsp nbsp testgettokenforcerefresh nbsp nbsp testgettingidtwicematches nbsp nbsp testgettingtokentwicematches messaging remote config failed tests nbsp nbsp nbsp nbsp testgetall storage failed tests nbsp nbsp testcreateworkingfolder nbsp nbsp testdeletefile nbsp nbsp testdownloadurl nbsp nbsp testinvalidatingreferenceswhendeletingapp nbsp nbsp testinvalidatingreferenceswhendeletingstorage nbsp nbsp testlargefilecancelupload nbsp nbsp testlargefilepauseresumeanddownloadcancel nbsp nbsp testputfileandgetfile nbsp nbsp testsignin nbsp nbsp teststorageurl nbsp nbsp testwriteandreadbytebuffer nbsp nbsp testwriteandreadfilewithcustommetadata | 1 |
201,480 | 15,209,044,840 | IssuesEvent | 2021-02-17 04:17:37 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] SimpleSecurityNetty4ServerTransportTests.testThreadContext | :Distributed/Network >test-failure Team:Distributed | **Build scan**: https://gradle-enterprise.elastic.co/s/xxazk5c2csshk
**Repro line**:
```
./gradlew ':x-pack:plugin:security:test' --tests "org.elasticsearch.xpack.security.transport.netty4.SimpleSecurityNetty4ServerTransportTests.testThreadContext" -Dtests.seed=3DD775158F20CC58 -Dtests.security.manager=true -Dtests.locale=tr -Dtests.timezone=Etc/GMT-0 -Druntime.java=11
```
**Reproduces locally?**: No
**Applicable branches**: master
**Failure history**:
**Failure excerpt**:
```
java.lang.AssertionError: this header is only visible in the handler context expected null, but was:<booooom>
at __randomizedtesting.SeedInfo.seed([3DD775158F20CC58:ABD4401512CF7604]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotNull(Assert.java:755)
at org.junit.Assert.assertNull(Assert.java:737)
at org.elasticsearch.transport.AbstractSimpleTransportTestCase.testThreadContext(AbstractSimpleTransportTestCase.java:383)
```
| 1.0 | [CI] SimpleSecurityNetty4ServerTransportTests.testThreadContext - **Build scan**: https://gradle-enterprise.elastic.co/s/xxazk5c2csshk
**Repro line**:
```
./gradlew ':x-pack:plugin:security:test' --tests "org.elasticsearch.xpack.security.transport.netty4.SimpleSecurityNetty4ServerTransportTests.testThreadContext" -Dtests.seed=3DD775158F20CC58 -Dtests.security.manager=true -Dtests.locale=tr -Dtests.timezone=Etc/GMT-0 -Druntime.java=11
```
**Reproduces locally?**: No
**Applicable branches**: master
**Failure history**:
**Failure excerpt**:
```
java.lang.AssertionError: this header is only visible in the handler context expected null, but was:<booooom>
at __randomizedtesting.SeedInfo.seed([3DD775158F20CC58:ABD4401512CF7604]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotNull(Assert.java:755)
at org.junit.Assert.assertNull(Assert.java:737)
at org.elasticsearch.transport.AbstractSimpleTransportTestCase.testThreadContext(AbstractSimpleTransportTestCase.java:383)
```
| test | testthreadcontext build scan repro line gradlew x pack plugin security test tests org elasticsearch xpack security transport testthreadcontext dtests seed dtests security manager true dtests locale tr dtests timezone etc gmt druntime java reproduces locally no applicable branches master failure history failure excerpt java lang assertionerror this header is only visible in the handler context expected null but was at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotnull assert java at org junit assert assertnull assert java at org elasticsearch transport abstractsimpletransporttestcase testthreadcontext abstractsimpletransporttestcase java | 1 |
127,717 | 10,479,773,991 | IssuesEvent | 2019-09-24 05:35:59 | linewalks/MDwalks-UI | https://api.github.com/repos/linewalks/MDwalks-UI | closed | chartUtility Test 만들기 | DONE test | **개요 및 요청사항**
- [x] strIdConvert
- [x] tableHeaderConvert
- [x] renderSVG
- [x] generateGroup
- [x] getStartAndEndTime
- [x] circleDataFilter
- [x] rectDataFilter
- [x] labelList
**결과물**
**완료 기준**
- [ ] PR Review @
| 1.0 | chartUtility Test 만들기 - **개요 및 요청사항**
- [x] strIdConvert
- [x] tableHeaderConvert
- [x] renderSVG
- [x] generateGroup
- [x] getStartAndEndTime
- [x] circleDataFilter
- [x] rectDataFilter
- [x] labelList
**결과물**
**완료 기준**
- [ ] PR Review @
| test | chartutility test 만들기 개요 및 요청사항 stridconvert tableheaderconvert rendersvg generategroup getstartandendtime circledatafilter rectdatafilter labellist 결과물 완료 기준 pr review | 1 |
95,649 | 8,569,239,211 | IssuesEvent | 2018-11-11 08:18:41 | IMA-WorldHealth/bhima-2.X | https://api.github.com/repos/IMA-WorldHealth/bhima-2.X | closed | barcodes.reverseLookup() is broken | Bug needs tests | Looks like we forgot to remove the .done() calls in the controllers/reports.js. Since we use async/await and `.done()` is nonstandard, the two are not compatible. We should makes sure we test our barcodes via integration testing. | 1.0 | barcodes.reverseLookup() is broken - Looks like we forgot to remove the .done() calls in the controllers/reports.js. Since we use async/await and `.done()` is nonstandard, the two are not compatible. We should makes sure we test our barcodes via integration testing. | test | barcodes reverselookup is broken looks like we forgot to remove the done calls in the controllers reports js since we use async await and done is nonstandard the two are not compatible we should makes sure we test our barcodes via integration testing | 1 |
112,266 | 4,514,861,102 | IssuesEvent | 2016-09-05 02:43:07 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | Facilitating checking sessions in approval | admin high priority next quick user interface | We want the tability to "tick off" annotations made by the community curator where we agree with their annotations. When they make a lot of annotations it can get a bit confusing otherwise.
One suggestion:
-highlight all of the annotation (turn it red or something) and then the ability to hit an admin only check box when we check the annotation off. | 1.0 | Facilitating checking sessions in approval - We want the tability to "tick off" annotations made by the community curator where we agree with their annotations. When they make a lot of annotations it can get a bit confusing otherwise.
One suggestion:
-highlight all of the annotation (turn it red or something) and then the ability to hit an admin only check box when we check the annotation off. | non_test | facilitating checking sessions in approval we want the tability to tick off annotations made by the community curator where we agree with their annotations when they make a lot of annotations it can get a bit confusing otherwise one suggestion highlight all of the annotation turn it red or something and then the ability to hit an admin only check box when we check the annotation off | 0 |
16,479 | 2,615,116,951 | IssuesEvent | 2015-03-01 05:42:09 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | calendar-v2-jsonc-oauth-sample | auto-migrated Priority-Low Type-Sample | ```
Which API and version (e.g. Google Calendar Data API version 2)?
Google Calendar Data API version 2
What format (e.g. JSON, Atom)?
JSON-C
What Authentation (e.g. OAuth, OAuth 2, Android, ClientLogin)?
OAuth
Java environment (e.g. Java 6, Android 2.2, App Engine 1.3.7)?
Java 6
External references, such as API reference guide?
http://code.google.com/apis/calendar/data/2.0/developers_guide_protocol.html
Please provide any additional information below.
```
Original issue reported on code.google.com by `yan...@google.com` on 7 Oct 2010 at 3:07 | 1.0 | calendar-v2-jsonc-oauth-sample - ```
Which API and version (e.g. Google Calendar Data API version 2)?
Google Calendar Data API version 2
What format (e.g. JSON, Atom)?
JSON-C
What Authentation (e.g. OAuth, OAuth 2, Android, ClientLogin)?
OAuth
Java environment (e.g. Java 6, Android 2.2, App Engine 1.3.7)?
Java 6
External references, such as API reference guide?
http://code.google.com/apis/calendar/data/2.0/developers_guide_protocol.html
Please provide any additional information below.
```
Original issue reported on code.google.com by `yan...@google.com` on 7 Oct 2010 at 3:07 | non_test | calendar jsonc oauth sample which api and version e g google calendar data api version google calendar data api version what format e g json atom json c what authentation e g oauth oauth android clientlogin oauth java environment e g java android app engine java external references such as api reference guide please provide any additional information below original issue reported on code google com by yan google com on oct at | 0 |
31,070 | 11,867,197,766 | IssuesEvent | 2020-03-26 06:17:02 | diecutfacemasks/diecutfacemasks.github.io | https://api.github.com/repos/diecutfacemasks/diecutfacemasks.github.io | closed | Fix TLS for diecutfacemasks.org | bug security ux | https://diecutfacemasks.org creates a certificate warning, this needs to be fixed to ensure user friendly and secure access to the web site. | True | Fix TLS for diecutfacemasks.org - https://diecutfacemasks.org creates a certificate warning, this needs to be fixed to ensure user friendly and secure access to the web site. | non_test | fix tls for diecutfacemasks org creates a certificate warning this needs to be fixed to ensure user friendly and secure access to the web site | 0 |
6,327 | 3,011,040,529 | IssuesEvent | 2015-07-28 15:57:07 | softlayer/sl-ember-components | https://api.github.com/repos/softlayer/sl-ember-components | closed | Confirm whether sl-panel has a dependency on Bootstrap-Datepicker | 2 - Working documentation research | The `#/browsers` page in the demo app indicates as such but I'm thinking this is not accurate. If it is not update the compatibility table.
<!---
@huboard:{"milestone_order":364.9999999925494}
-->
| 1.0 | Confirm whether sl-panel has a dependency on Bootstrap-Datepicker - The `#/browsers` page in the demo app indicates as such but I'm thinking this is not accurate. If it is not update the compatibility table.
<!---
@huboard:{"milestone_order":364.9999999925494}
-->
| non_test | confirm whether sl panel has a dependency on bootstrap datepicker the browsers page in the demo app indicates as such but i m thinking this is not accurate if it is not update the compatibility table huboard milestone order | 0 |
340,365 | 10,271,233,953 | IssuesEvent | 2019-08-23 13:40:26 | infor-design/enterprise | https://api.github.com/repos/infor-design/enterprise | closed | Editor: Event callback for mode change - Visual to HTML & vice versa | [3] priority: high type: enhancement :sparkles: | **Description**
We need to have event callback when user changes mode from HTML to Visual or vice versa.
We would like to do some transformations on HTML code before it is shown in the Visual mode. We would also like to remove those transformations when user switches back to HTML code. To support this, we would like to have call backs where we can listen to mode change events.
For example,
User writes following in HTML mode
```
<p>
Hello {{customeName}},
Thank you for ordering.
{{#if selectedForSurvey }}
We would like to hear your feedback. Please fill up this survey form.
{{/if}}
</p>
```
When he switches to Visual mode, we would like to show the user with actual values for variables used in HTML code. To perform these transformations, we would like to have some callbacks to listen to mode change events.
We need this in our Angular project. So `enterprise-ng` also needs the fix.
**Possible Solution**
We need some callback functions like below. The name suggested are for demonstration purpose only. You can choose whatever names are more suitable for
```
this.editor.onBeforePreview(
() => {
this.editorElementRef.nativeHTML = transform(this.editorElementRef.nativeHTML);
}
);
```
```
this.editor.onBeforeHTML(
() => {
this.editorElementRef.nativeHTML = reverseTransform(this.editorElementRef.nativeHTML);
}
);
```
**Additional context**
We would appreciate if this issue can fixed on a high priority basis. We won't be able to show proper preview to our customers without supporting the transformations on HTML code. | 1.0 | Editor: Event callback for mode change - Visual to HTML & vice versa - **Description**
We need to have event callback when user changes mode from HTML to Visual or vice versa.
We would like to do some transformations on HTML code before it is shown in the Visual mode. We would also like to remove those transformations when user switches back to HTML code. To support this, we would like to have call backs where we can listen to mode change events.
For example,
User writes following in HTML mode
```
<p>
Hello {{customeName}},
Thank you for ordering.
{{#if selectedForSurvey }}
We would like to hear your feedback. Please fill up this survey form.
{{/if}}
</p>
```
When he switches to Visual mode, we would like to show the user with actual values for variables used in HTML code. To perform these transformations, we would like to have some callbacks to listen to mode change events.
We need this in our Angular project. So `enterprise-ng` also needs the fix.
**Possible Solution**
We need some callback functions like below. The name suggested are for demonstration purpose only. You can choose whatever names are more suitable for
```
this.editor.onBeforePreview(
() => {
this.editorElementRef.nativeHTML = transform(this.editorElementRef.nativeHTML);
}
);
```
```
this.editor.onBeforeHTML(
() => {
this.editorElementRef.nativeHTML = reverseTransform(this.editorElementRef.nativeHTML);
}
);
```
**Additional context**
We would appreciate if this issue can fixed on a high priority basis. We won't be able to show proper preview to our customers without supporting the transformations on HTML code. | non_test | editor event callback for mode change visual to html vice versa description we need to have event callback when user changes mode from html to visual or vice versa we would like to do some transformations on html code before it is shown in the visual mode we would also like to remove those transformations when user switches back to html code to support this we would like to have call backs where we can listen to mode change events for example user writes following in html mode hello customename thank you for ordering if selectedforsurvey we would like to hear your feedback please fill up this survey form if when he switches to visual mode we would like to show the user with actual values for variables used in html code to perform these transformations we would like to have some callbacks to listen to mode change events we need this in our angular project so enterprise ng also needs the fix possible solution we need some callback functions like below the name suggested are for demonstration purpose only you can choose whatever names are more suitable for this editor onbeforepreview this editorelementref nativehtml transform this editorelementref nativehtml this editor onbeforehtml this editorelementref nativehtml reversetransform this editorelementref nativehtml additional context we would appreciate if this issue can fixed on a high priority basis we won t be able to show proper preview to our customers without supporting the transformations on html code | 0 |
528,548 | 15,369,450,739 | IssuesEvent | 2021-03-02 07:22:57 | urbit/landscape | https://api.github.com/repos/urbit/landscape | closed | chat: mention punctuation handling renders !~sampel-palnet as ~sampel-palnet! | priority: 3 | **Describe the bug**
punctuation preffixed to a mention is rendered suffixed to the mention
`!!!!!!!!!~sampel-palnet` becomes `~sampel-palnet!!!!!!!!!` and
`!!!~sampel-palnet;;;` becomes `~sampel-palnet!!!;;;`
**To Reproduce**
Steps to reproduce the behavior:
1. prefix punctuation to mention
2. observe said punctuation now suffixed
**Expected behavior**
punctuation should stay where it belongs
**Desktop (please complete the following information):**
- OS: arch linux
- Browser qutebrowser
- Base hash eljue | 1.0 | chat: mention punctuation handling renders !~sampel-palnet as ~sampel-palnet! - **Describe the bug**
punctuation preffixed to a mention is rendered suffixed to the mention
`!!!!!!!!!~sampel-palnet` becomes `~sampel-palnet!!!!!!!!!` and
`!!!~sampel-palnet;;;` becomes `~sampel-palnet!!!;;;`
**To Reproduce**
Steps to reproduce the behavior:
1. prefix punctuation to mention
2. observe said punctuation now suffixed
**Expected behavior**
punctuation should stay where it belongs
**Desktop (please complete the following information):**
- OS: arch linux
- Browser qutebrowser
- Base hash eljue | non_test | chat mention punctuation handling renders sampel palnet as sampel palnet describe the bug punctuation preffixed to a mention is rendered suffixed to the mention sampel palnet becomes sampel palnet and sampel palnet becomes sampel palnet to reproduce steps to reproduce the behavior prefix punctuation to mention observe said punctuation now suffixed expected behavior punctuation should stay where it belongs desktop please complete the following information os arch linux browser qutebrowser base hash eljue | 0 |
740,196 | 25,740,173,244 | IssuesEvent | 2022-12-08 05:14:20 | googleapis/nodejs-ai-platform | https://api.github.com/repos/googleapis/nodejs-ai-platform | closed | AI platform get training pipeline: should get the specified training pipeline failed | type: bug priority: p1 flakybot: issue api: vertex-ai | Note: #434 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined
7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended.
Error: Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined
7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-training-pipeline.test.js:23:28)
at Context.<anonymous> (test/get-training-pipeline.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | 1.0 | AI platform get training pipeline: should get the specified training pipeline failed - Note: #434 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined
7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended.
Error: Command failed: node ./get-training-pipeline.js 1419759782528548864 undefined undefined
7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-training-pipeline.test.js:23:28)
at Context.<anonymous> (test/get-training-pipeline.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | non_test | ai platform get training pipeline should get the specified training pipeline failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output command failed node get training pipeline js undefined undefined permission denied permission denied consumer project undefined has been suspended error command failed node get training pipeline js undefined undefined permission denied permission denied consumer project undefined has been suspended at checkexecsyncerror child process js at object execsync child process js at execsync test get training pipeline test js at context test get training pipeline test js at processimmediate internal timers js | 0 |
17,592 | 3,012,747,540 | IssuesEvent | 2015-07-29 02:09:31 | yawlfoundation/yawl | https://api.github.com/repos/yawlfoundation/yawl | closed | [CLOSED] Case start fails | auto-migrated Priority-Critical Type-Defect | <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:21 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/41_
----
```
Starting a case for the attached example yields the following error
message: "Unsuccessful case start:Unexpected failure from launchCase (see
log for details".
Task A has the RRR strategy, the other task is a timer task.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 22 Jul 2008 at 8:06
Attachments:
* [new3.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-41/comment-0/new3.xml)
| 1.0 | [CLOSED] Case start fails - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:21 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/41_
----
```
Starting a case for the attached example yields the following error
message: "Unsuccessful case start:Unexpected failure from launchCase (see
log for details".
Task A has the RRR strategy, the other task is a timer task.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 22 Jul 2008 at 8:06
Attachments:
* [new3.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-41/comment-0/new3.xml)
| non_test | case start fails issue by monday jul at gmt originally opened as starting a case for the attached example yields the following error message unsuccessful case start unexpected failure from launchcase see log for details task a has the rrr strategy the other task is a timer task original issue reported on code google com by arthurte gmail com on jul at attachments | 0 |
343,713 | 30,685,508,790 | IssuesEvent | 2023-07-26 12:06:03 | sarahsporck/Migration-Test-Repository | https://api.github.com/repos/sarahsporck/Migration-Test-Repository | closed | IGAPP-240: Test module wohnen | Native Test |
# Max Ammann - 19.3.2019, 18:16:09
None
**Environment**: -
**Linked issues:**
| 1.0 | IGAPP-240: Test module wohnen -
# Max Ammann - 19.3.2019, 18:16:09
None
**Environment**: -
**Linked issues:**
| test | igapp test module wohnen max ammann none environment linked issues | 1 |
204,061 | 15,398,713,506 | IssuesEvent | 2021-03-04 00:36:50 | nucleus-security/Test-repo | https://api.github.com/repos/nucleus-security/Test-repo | opened | Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [High] - CentOS Security Update for kernel Security Update (CESA-2019:1169) | Testing | Source: QUALYS
Finding Description: CentOS has released security update for kernel security update to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6
Impact: N/A</p>
Target(s): Asset name: 45.55.254.143
IP: 45.55.254.143
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">centos 6</a> for updates and patch information.
<p>Patch:<br />
Following are links for downloading patches to fix the vulnerabilities:
</p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">CESA-2019:1169: centos 6</a></p>
References:
ID:256612
CVE:CVE-2018-12130,CVE-2018-12126,CVE-2018-12127,CVE-2019-11091
Category:CentOS
PCI Flagged:1
Vendor References:CESA-2019:1169 centos 6
Severity: High
Date Discovered: 2020-01-07 14:35:48
Nucleus Notification Rules Triggered: GitHub Rule
Project Name: Ticketing Rules now apply to all vulnerabilities
| 1.0 | Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [High] - CentOS Security Update for kernel Security Update (CESA-2019:1169) - Source: QUALYS
Finding Description: CentOS has released security update for kernel security update to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6
Impact: N/A</p>
Target(s): Asset name: 45.55.254.143
IP: 45.55.254.143
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">centos 6</a> for updates and patch information.
<p>Patch:<br />
Following are links for downloading patches to fix the vulnerabilities:
</p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2019-may/023309.html">CESA-2019:1169: centos 6</a></p>
References:
ID:256612
CVE:CVE-2018-12130,CVE-2018-12126,CVE-2018-12127,CVE-2019-11091
Category:CentOS
PCI Flagged:1
Vendor References:CESA-2019:1169 centos 6
Severity: High
Date Discovered: 2020-01-07 14:35:48
Nucleus Notification Rules Triggered: GitHub Rule
Project Name: Ticketing Rules now apply to all vulnerabilities
| test | nucleus project ticketing rules now apply to all vulnerabilities centos security update for kernel security update cesa source qualys finding description centos has released security update for kernel security update to fix the vulnerabilities affected products centos impact n a target s asset name ip solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory for updates and patch information patch following are links for downloading patches to fix the vulnerabilities references id cve cve cve cve cve category centos pci flagged vendor references cesa centos severity high date discovered nucleus notification rules triggered github rule project name ticketing rules now apply to all vulnerabilities | 1 |
113,054 | 24,356,999,738 | IssuesEvent | 2022-10-03 08:22:46 | julz0815/test-action | https://api.github.com/repos/julz0815/test-action | opened | Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection') [VID:7] | VeracodeFlaw: Very High Veracode Policy Scan | https://github.com/julz0815/test-action/blob/1b75779c50b856e0bd6418d47c0aa10cfa479423/services/admin.service.js#L1-L11
**Filename:** admin.service.js
**Line:** 6
**CWE:** 78 (Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection'))
<span>This call to util.exec() contains a command injection flaw. The argument to the function is constructed using untrusted input. If an attacker is allowed to specify all or part of the command, it may be possible to execute commands on the server with the privileges of the executing process. The level of exposure depends on the effectiveness of input validation routines, if any. </span> <span>Validate all untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible. When using blocklists, be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters. Most APIs that execute system commands also have a "safe" version of the method that takes an array of strings as input rather than a single string, which protects against some forms of command injection.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/78.html">CWE</a> <a href="https://owasp.org/www-community/attacks/Command_Injection">OWASP</a></span> | 2.0 | Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection') [VID:7] - https://github.com/julz0815/test-action/blob/1b75779c50b856e0bd6418d47c0aa10cfa479423/services/admin.service.js#L1-L11
**Filename:** admin.service.js
**Line:** 6
**CWE:** 78 (Improper Neutralization of Special Elements used in an OS Command ('OS Command Injection') ('Command or Argument Injection'))
<span>This call to util.exec() contains a command injection flaw. The argument to the function is constructed using untrusted input. If an attacker is allowed to specify all or part of the command, it may be possible to execute commands on the server with the privileges of the executing process. The level of exposure depends on the effectiveness of input validation routines, if any. </span> <span>Validate all untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible. When using blocklists, be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters. Most APIs that execute system commands also have a "safe" version of the method that takes an array of strings as input rather than a single string, which protects against some forms of command injection.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/78.html">CWE</a> <a href="https://owasp.org/www-community/attacks/Command_Injection">OWASP</a></span> | non_test | improper neutralization of special elements used in an os command os command injection command or argument injection filename admin service js line cwe improper neutralization of special elements used in an os command os command injection command or argument injection this call to util exec contains a command injection flaw the argument to the function is constructed using untrusted input if an attacker is allowed to specify all or part of the command it may be possible to execute commands on the server with the privileges of the executing process the level of exposure depends on the effectiveness of input validation routines if any validate all untrusted input to ensure that it conforms to the expected format using centralized data validation routines when possible when using blocklists be sure that the sanitizing routine performs a sufficient number of iterations to remove all instances of disallowed characters most apis that execute system commands also have a safe version of the method that takes an array of strings as input rather than a single string which protects against some forms of command injection references a href a href | 0 |
660,197 | 21,956,727,160 | IssuesEvent | 2022-05-24 12:45:49 | EddieHubCommunity/LinkFree | https://api.github.com/repos/EddieHubCommunity/LinkFree | closed | [BUG] Copy to clipboard in the Share to social media icons is not working | 🕹 aspect: interface 🛠 goal: fix 🟧 priority: high | ### Description
The following button is not working in the `share to social media`
- copy to clip board
I tried clicking the button few times and it does not seem to copy the contents of the link. Instead it selects the contents of the description in the profile.
I am using BRAVE BROWSER.
At the time of this issue the linkfree is at the v0.58.3
### Screenshots

### Additional information
_No response_ | 1.0 | [BUG] Copy to clipboard in the Share to social media icons is not working - ### Description
The following button is not working in the `share to social media`
- copy to clip board
I tried clicking the button few times and it does not seem to copy the contents of the link. Instead it selects the contents of the description in the profile.
I am using BRAVE BROWSER.
At the time of this issue the linkfree is at the v0.58.3
### Screenshots

### Additional information
_No response_ | non_test | copy to clipboard in the share to social media icons is not working description the following button is not working in the share to social media copy to clip board i tried clicking the button few times and it does not seem to copy the contents of the link instead it selects the contents of the description in the profile i am using brave browser at the time of this issue the linkfree is at the screenshots additional information no response | 0 |
14,484 | 3,407,935,839 | IssuesEvent | 2015-12-04 07:32:27 | thedaydreamer/SampleRepo | https://api.github.com/repos/thedaydreamer/SampleRepo | opened | Awesome People | ZipTest | Awesome people are always awesome
1. point one
2. point two
You can view a snapshot of the issue using following link :
http://webconverger.org/img/lollipop.png | 1.0 | Awesome People - Awesome people are always awesome
1. point one
2. point two
You can view a snapshot of the issue using following link :
http://webconverger.org/img/lollipop.png | test | awesome people awesome people are always awesome point one point two you can view a snapshot of the issue using following link | 1 |
338,770 | 30,320,405,811 | IssuesEvent | 2023-07-10 18:46:32 | ita-social-projects/StreetCode | https://api.github.com/repos/ita-social-projects/StreetCode | opened | Admin [Interesting facts block] Verify if Admin can add no more than 62 symbols to title field | Test case | **Priority:*hight*
**Description**:
Test case verifies if Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working
**Preconditions**
1. Go to http://185.230.138.173/admin-panel
2. Login as 'adminStreetcode ' , password - 'pH2603VkN4d'
3. Go to the Interesting facts block - WOW - факти
**Steps to reproduce**
1. Click on "+" button
2. Add the text to the field "Заголовок"
3. Verify if Counter for "symbols left" is displayed and correctly counts all symbols in the text including spaces
4. Add the text with more than 62 symbols to the field "Заголовок"
5. Сheck that admin cannot add text longer than 62 characters in this field, and this title field only allows to add 62 characters.
**Expected result**
Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working
**User story:**
#123
**Related bug:**
| 1.0 | Admin [Interesting facts block] Verify if Admin can add no more than 62 symbols to title field - **Priority:*hight*
**Description**:
Test case verifies if Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working
**Preconditions**
1. Go to http://185.230.138.173/admin-panel
2. Login as 'adminStreetcode ' , password - 'pH2603VkN4d'
3. Go to the Interesting facts block - WOW - факти
**Steps to reproduce**
1. Click on "+" button
2. Add the text to the field "Заголовок"
3. Verify if Counter for "symbols left" is displayed and correctly counts all symbols in the text including spaces
4. Add the text with more than 62 symbols to the field "Заголовок"
5. Сheck that admin cannot add text longer than 62 characters in this field, and this title field only allows to add 62 characters.
**Expected result**
Admin can add no more than 62 symbols to title field and Counter for "symbols left" is displayed and working
**User story:**
#123
**Related bug:**
| test | admin verify if admin can add no more than symbols to title field priority hight description test case verifies if admin can add no more than symbols to title field and counter for symbols left is displayed and working preconditions go to login as adminstreetcode password go to the interesting facts block wow факти steps to reproduce click on button add the text to the field заголовок verify if counter for symbols left is displayed and correctly counts all symbols in the text including spaces add the text with more than symbols to the field заголовок сheck that admin cannot add text longer than characters in this field and this title field only allows to add characters expected result admin can add no more than symbols to title field and counter for symbols left is displayed and working user story related bug | 1 |
62,690 | 6,803,385,674 | IssuesEvent | 2017-11-03 00:36:32 | NetsBlox/NetsBlox | https://api.github.com/repos/NetsBlox/NetsBlox | closed | newly created projects (programmatically) cannot be published | bug minor testing | If you create a project programmatically, it cannot be saved as it relies on accessing a non-standard field only available when reading from the database (deprecated and should not be used). This is also a problem exposed by #1637 | 1.0 | newly created projects (programmatically) cannot be published - If you create a project programmatically, it cannot be saved as it relies on accessing a non-standard field only available when reading from the database (deprecated and should not be used). This is also a problem exposed by #1637 | test | newly created projects programmatically cannot be published if you create a project programmatically it cannot be saved as it relies on accessing a non standard field only available when reading from the database deprecated and should not be used this is also a problem exposed by | 1 |
117,120 | 17,410,187,498 | IssuesEvent | 2021-08-03 11:18:00 | tamirdahan/keycloak | https://api.github.com/repos/tamirdahan/keycloak | closed | WS-2017-0116 (Medium) detected in angular-v1.2.32 - autoclosed | security vulnerability | ## WS-2017-0116 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-v1.2.32</b></p></summary>
<p></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json</p>
<p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower_components/angular/.bower.json</p>
<p>
Dependency Hierarchy:
- :x: **angular-v1.2.32** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The use element can reference external svg's (same origin) and can include xlink javascript urls or foreign object that can execute xss.
<p>Publish Date: 2015-12-05
<p>URL: <a href=https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8>WS-2017-0116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/pull/13453">https://github.com/angular/angular.js/pull/13453</a></p>
<p>Release Date: 2015-12-05</p>
<p>Fix Resolution: angular - 1.5.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Bower","packageName":"angular","packageVersion":"v1.2.32","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json"],"isTransitiveDependency":false,"dependencyTree":"angular:v1.2.32","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0116","vulnerabilityDetails":"The use element can reference external svg\u0027s (same origin) and can include xlink javascript urls or foreign object that can execute xss.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8","cvss3Severity":"medium","cvss3Score":"5.8","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2017-0116 (Medium) detected in angular-v1.2.32 - autoclosed - ## WS-2017-0116 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-v1.2.32</b></p></summary>
<p></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json</p>
<p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower_components/angular/.bower.json</p>
<p>
Dependency Hierarchy:
- :x: **angular-v1.2.32** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The use element can reference external svg's (same origin) and can include xlink javascript urls or foreign object that can execute xss.
<p>Publish Date: 2015-12-05
<p>URL: <a href=https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8>WS-2017-0116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/pull/13453">https://github.com/angular/angular.js/pull/13453</a></p>
<p>Release Date: 2015-12-05</p>
<p>Fix Resolution: angular - 1.5.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Bower","packageName":"angular","packageVersion":"v1.2.32","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/autofill-event/bower.json"],"isTransitiveDependency":false,"dependencyTree":"angular:v1.2.32","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0116","vulnerabilityDetails":"The use element can reference external svg\u0027s (same origin) and can include xlink javascript urls or foreign object that can execute xss.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/7a668cdd7d08a7016883eb3c671cbcd586223ae8","cvss3Severity":"medium","cvss3Score":"5.8","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_test | ws medium detected in angular autoclosed ws medium severity vulnerability vulnerable library angular path to dependency file keycloak themes src main resources theme keycloak common resources node modules autofill event bower json path to vulnerable library keycloak themes src main resources theme keycloak common resources node modules autofill event bower components angular bower json dependency hierarchy x angular vulnerable library found in base branch master vulnerability details the use element can reference external svg s same origin and can include xlink javascript urls or foreign object that can execute xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angular isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree angular isminimumfixversionavailable true minimumfixversion angular basebranches vulnerabilityidentifier ws vulnerabilitydetails the use element can reference external svg same origin and can include xlink javascript urls or foreign object that can execute xss vulnerabilityurl | 0 |
21,597 | 10,666,994,584 | IssuesEvent | 2019-10-19 08:49:01 | OSWeekends/guilds.osweekends.com | https://api.github.com/repos/OSWeekends/guilds.osweekends.com | opened | WS-2019-0064 (High) detected in handlebars-4.0.11.tgz | security vulnerability | ## WS-2019-0064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/guilds.osweekends.com/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/guilds.osweekends.com/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- pillars-0.7.1.tgz (Root Library)
- templated-0.3.9.tgz
- :x: **handlebars-4.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/OSWeekends/guilds.osweekends.com/commit/7c3567c7b8d78dce0fcb947eca5db3cca195eb19">7c3567c7b8d78dce0fcb947eca5db3cca195eb19</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0064 (High) detected in handlebars-4.0.11.tgz - ## WS-2019-0064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/guilds.osweekends.com/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/guilds.osweekends.com/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- pillars-0.7.1.tgz (Root Library)
- templated-0.3.9.tgz
- :x: **handlebars-4.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/OSWeekends/guilds.osweekends.com/commit/7c3567c7b8d78dce0fcb947eca5db3cca195eb19">7c3567c7b8d78dce0fcb947eca5db3cca195eb19</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws high detected in handlebars tgz ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm guilds osweekends com package json path to vulnerable library tmp ws scm guilds osweekends com node modules handlebars package json dependency hierarchy pillars tgz root library templated tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details versions of handlebars prior to are vulnerable to prototype pollution templates may alter an objects prototype thus allowing an attacker to execute arbitrary code on the server publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
154,105 | 13,537,449,117 | IssuesEvent | 2020-09-16 10:31:08 | jolocom/jolo-did-method | https://api.github.com/repos/jolocom/jolo-did-method | closed | Improve documentation | documentation | - Briefly outline the contents of the `packges` folder on the top level of the repo, including references to how and where the separate modules are used.
- Document the individual packages. | 1.0 | Improve documentation - - Briefly outline the contents of the `packges` folder on the top level of the repo, including references to how and where the separate modules are used.
- Document the individual packages. | non_test | improve documentation briefly outline the contents of the packges folder on the top level of the repo including references to how and where the separate modules are used document the individual packages | 0 |
160,147 | 12,504,901,263 | IssuesEvent | 2020-06-02 09:46:50 | Oldes/Rebol-issues | https://api.github.com/repos/Oldes/Rebol-issues | closed | Save/encode png fails under unix platforms | Test.written Type.bug | _Submitted by:_ **johnk**
The following works fine under Windows, but fails on Linux x86, Max OS X and Linux ARM (raspberry pi).
It seems to be related to encode.
Saving to a bmp works correctly as expected.
``` rebol
save %test.png make image! [1024x768]
encode 'png make image! [1024x768]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2040)** [ Version: r3 master Type: Bug Platform: Linux x86 libc6 Category: Unspecified Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2040</sup>
Comments:
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **abolka**
Also crashes for me on Win32 (using http://www.rebolsource.net/downloads/win32-x86/r3-g4d9840f.exe).
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **fork**
I believe this to be resolved by https://github.com/rebol/rebol/pull/155
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **abolka**
In the core-tests suite.
---
> **Rebolbot** commented on Mar 2, 2014:
_Submitted by:_ **abolka**
Also addressed by https://github.com/rebol/rebol/pull/194.
---
> **Rebolbot** added **Type.bug** and **Test.written** on Jan 12, 2016
---
> **Oldes** commented on May 29, 2019:
This one should be closed as it is not an issue anymore.
--- | 1.0 | Save/encode png fails under unix platforms - _Submitted by:_ **johnk**
The following works fine under Windows, but fails on Linux x86, Max OS X and Linux ARM (raspberry pi).
It seems to be related to encode.
Saving to a bmp works correctly as expected.
``` rebol
save %test.png make image! [1024x768]
encode 'png make image! [1024x768]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2040)** [ Version: r3 master Type: Bug Platform: Linux x86 libc6 Category: Unspecified Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2040</sup>
Comments:
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **abolka**
Also crashes for me on Win32 (using http://www.rebolsource.net/downloads/win32-x86/r3-g4d9840f.exe).
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **fork**
I believe this to be resolved by https://github.com/rebol/rebol/pull/155
---
> **Rebolbot** commented on Sep 30, 2013:
_Submitted by:_ **abolka**
In the core-tests suite.
---
> **Rebolbot** commented on Mar 2, 2014:
_Submitted by:_ **abolka**
Also addressed by https://github.com/rebol/rebol/pull/194.
---
> **Rebolbot** added **Type.bug** and **Test.written** on Jan 12, 2016
---
> **Oldes** commented on May 29, 2019:
This one should be closed as it is not an issue anymore.
--- | test | save encode png fails under unix platforms submitted by johnk the following works fine under windows but fails on linux max os x and linux arm raspberry pi it seems to be related to encode saving to a bmp works correctly as expected rebol save test png make image encode png make image imported from imported from comments rebolbot commented on sep submitted by abolka also crashes for me on using rebolbot commented on sep submitted by fork i believe this to be resolved by rebolbot commented on sep submitted by abolka in the core tests suite rebolbot commented on mar submitted by abolka also addressed by rebolbot added type bug and test written on jan oldes commented on may this one should be closed as it is not an issue anymore | 1 |
91,693 | 10,726,867,634 | IssuesEvent | 2019-10-28 10:20:45 | suvajit-sarkar/github-upload | https://api.github.com/repos/suvajit-sarkar/github-upload | opened | Indy: Ansible role to get EIP metadata from AWS | documentation | Prepare an Ansible role which retrieves the EIP allocation IDs for all steward public IPs found in the network.yaml and stores them in an Ansible fact / variable to be used later. The role should also check that the same public IP is specified for all stewards within one organization. | 1.0 | Indy: Ansible role to get EIP metadata from AWS - Prepare an Ansible role which retrieves the EIP allocation IDs for all steward public IPs found in the network.yaml and stores them in an Ansible fact / variable to be used later. The role should also check that the same public IP is specified for all stewards within one organization. | non_test | indy ansible role to get eip metadata from aws prepare an ansible role which retrieves the eip allocation ids for all steward public ips found in the network yaml and stores them in an ansible fact variable to be used later the role should also check that the same public ip is specified for all stewards within one organization | 0 |
528,145 | 15,360,844,578 | IssuesEvent | 2021-03-01 17:24:16 | edgedb/edgedb | https://api.github.com/repos/edgedb/edgedb | closed | Migrations aren't propagated to existing connections | bug high priority | - EdgeDB Version:1.0-alpha.8+dev.5386.g6933b454f
Steps to Reproduce:
1. Start `edgedb` (repl) and do a query:
```
edgedb> SELECT Counter;
error: object type or alias 'default::Counter' does not exist
┌─ query:1:8
│
1 │ SELECT Counter;
│ ^^^^^^^ error
```
2. Then in another shell do `edgedb migrate` to create the type
3. Then return back to this shell and confirm that error still there (while it shouldn't be):
```
edgedb> SELECT Counter;
error: object type or alias 'default::Counter' does not exist
┌─ query:1:8
│
1 │ SELECT Counter;
│ ^^^^^^^ error
```
4. Also try and confirm that it works in a new connection:
```
edgedb> SELECT Counter;
{}
```
**Note:** this doesn't reproduce with `CREATE TYPE` only with migrations. | 1.0 | Migrations aren't propagated to existing connections - - EdgeDB Version:1.0-alpha.8+dev.5386.g6933b454f
Steps to Reproduce:
1. Start `edgedb` (repl) and do a query:
```
edgedb> SELECT Counter;
error: object type or alias 'default::Counter' does not exist
┌─ query:1:8
│
1 │ SELECT Counter;
│ ^^^^^^^ error
```
2. Then in another shell do `edgedb migrate` to create the type
3. Then return back to this shell and confirm that error still there (while it shouldn't be):
```
edgedb> SELECT Counter;
error: object type or alias 'default::Counter' does not exist
┌─ query:1:8
│
1 │ SELECT Counter;
│ ^^^^^^^ error
```
4. Also try and confirm that it works in a new connection:
```
edgedb> SELECT Counter;
{}
```
**Note:** this doesn't reproduce with `CREATE TYPE` only with migrations. | non_test | migrations aren t propagated to existing connections edgedb version alpha dev steps to reproduce start edgedb repl and do a query edgedb select counter error object type or alias default counter does not exist ┌─ query │ │ select counter │ error then in another shell do edgedb migrate to create the type then return back to this shell and confirm that error still there while it shouldn t be edgedb select counter error object type or alias default counter does not exist ┌─ query │ │ select counter │ error also try and confirm that it works in a new connection edgedb select counter note this doesn t reproduce with create type only with migrations | 0 |
184,309 | 31,854,850,425 | IssuesEvent | 2023-09-15 06:33:54 | opengovsg/FormSG | https://api.github.com/repos/opengovsg/FormSG | closed | Announcement modal | design P4 | Why
- Inform users of major new features
- better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to
Qns
- Is the v3 "What's New" / notifications design sufficient?
Implementation
- Can check out Postman's https://github.com/opengovsg/postmangovsg/pull/920 | 1.0 | Announcement modal - Why
- Inform users of major new features
- better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to
Qns
- Is the v3 "What's New" / notifications design sufficient?
Implementation
- Can check out Postman's https://github.com/opengovsg/postmangovsg/pull/920 | non_test | announcement modal why inform users of major new features better than email as users can try it out straight away instead of the notification coming in between possibly other work they have to attend to qns is the what s new notifications design sufficient implementation can check out postman s | 0 |
12,341 | 3,266,862,989 | IssuesEvent | 2015-10-22 22:46:45 | NorthBridge/nexus-community | https://api.github.com/repos/NorthBridge/nexus-community | closed | Nexus Web Meet timeout oid query param value incorrect | ready to test | Should show a message that indicates user session has timed out. Right now message says they logged out. | 1.0 | Nexus Web Meet timeout oid query param value incorrect - Should show a message that indicates user session has timed out. Right now message says they logged out. | test | nexus web meet timeout oid query param value incorrect should show a message that indicates user session has timed out right now message says they logged out | 1 |
289,502 | 24,993,866,330 | IssuesEvent | 2022-11-02 21:27:25 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | acctest: Failing acceptance tests | tests | Acceptance tests are failing in the `acctest` package:
```
=== RUN TestAccProvider_DefaultTagsTags_none
=== PAUSE TestAccProvider_DefaultTagsTags_none
=== CONT TestAccProvider_DefaultTagsTags_none
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Starting TestCase: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Calling TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Called TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:14.038Z [WARN] sdk.helper_resource: Error running Terraform CLI command: test_terraform_path=/opt/teamcity-agent/temp/buildTmp/plugintest-terraform3209408513/terraform test_working_directory=/opt/teamcity-agent/temp/buildTmp/plugintest3846192360
error=
| exit status 1
|
| Error: Invalid data source
|
| on terraform_plugin_test.tf line 2, in data "aws_region" "provider_test":
| 2: data "aws_region" "provider_test" {}
|
| The provider hashicorp/aws does not support data source "aws_region".
|
| Error: Invalid data source
|
| on terraform_plugin_test.tf line 5, in data "aws_service" "provider_test":
| 5: data "aws_service" "provider_test" {
|
| The provider hashicorp/aws does not support data source "aws_service".
```
Relates https://github.com/hashicorp/terraform-provider-aws/pull/27221. | 1.0 | acctest: Failing acceptance tests - Acceptance tests are failing in the `acctest` package:
```
=== RUN TestAccProvider_DefaultTagsTags_none
=== PAUSE TestAccProvider_DefaultTagsTags_none
=== CONT TestAccProvider_DefaultTagsTags_none
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Starting TestCase: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Calling TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:13.462Z [DEBUG] sdk.helper_resource: Called TestCase PreCheck: test_name=TestAccProvider_DefaultTagsTags_none test_step_number=1
2022-11-02T05:10:14.038Z [WARN] sdk.helper_resource: Error running Terraform CLI command: test_terraform_path=/opt/teamcity-agent/temp/buildTmp/plugintest-terraform3209408513/terraform test_working_directory=/opt/teamcity-agent/temp/buildTmp/plugintest3846192360
error=
| exit status 1
|
| Error: Invalid data source
|
| on terraform_plugin_test.tf line 2, in data "aws_region" "provider_test":
| 2: data "aws_region" "provider_test" {}
|
| The provider hashicorp/aws does not support data source "aws_region".
|
| Error: Invalid data source
|
| on terraform_plugin_test.tf line 5, in data "aws_service" "provider_test":
| 5: data "aws_service" "provider_test" {
|
| The provider hashicorp/aws does not support data source "aws_service".
```
Relates https://github.com/hashicorp/terraform-provider-aws/pull/27221. | test | acctest failing acceptance tests acceptance tests are failing in the acctest package run testaccprovider defaulttagstags none pause testaccprovider defaulttagstags none cont testaccprovider defaulttagstags none sdk helper resource starting testcase test name testaccprovider defaulttagstags none test step number sdk helper resource calling testcase precheck test name testaccprovider defaulttagstags none test step number sdk helper resource called testcase precheck test name testaccprovider defaulttagstags none test step number sdk helper resource error running terraform cli command test terraform path opt teamcity agent temp buildtmp plugintest terraform test working directory opt teamcity agent temp buildtmp error exit status error invalid data source on terraform plugin test tf line in data aws region provider test data aws region provider test the provider hashicorp aws does not support data source aws region error invalid data source on terraform plugin test tf line in data aws service provider test data aws service provider test the provider hashicorp aws does not support data source aws service relates | 1 |
22,834 | 20,269,206,753 | IssuesEvent | 2022-02-15 14:50:01 | firemodels/fds | https://api.github.com/repos/firemodels/fds | closed | Output A and E in TGA_ANALYSIS mode | Usability | Simulations with TGA_ANALYSIS = .TRUE. usually start with REFERENCE_TEMPERATURE etc (first order), but if user wants to see the resulting A and E, and continue by specifying / modifying them directly, it would be convenient to see the calculated values from somewhere.
Normally, FDS writes A and E in .out file, under MATERIAL definitions. But TGA_ANALYSIS -mode does not produce .out file. Finding A and E requires removing TGA_ANALYSIS = .TRUE. -line, and making sure the file runs otherwise. Unnecessary steps in the process.
How about writing A and E directly to LU_ERR before the 'STOP: FDS performed.... ' -line? | True | Output A and E in TGA_ANALYSIS mode - Simulations with TGA_ANALYSIS = .TRUE. usually start with REFERENCE_TEMPERATURE etc (first order), but if user wants to see the resulting A and E, and continue by specifying / modifying them directly, it would be convenient to see the calculated values from somewhere.
Normally, FDS writes A and E in .out file, under MATERIAL definitions. But TGA_ANALYSIS -mode does not produce .out file. Finding A and E requires removing TGA_ANALYSIS = .TRUE. -line, and making sure the file runs otherwise. Unnecessary steps in the process.
How about writing A and E directly to LU_ERR before the 'STOP: FDS performed.... ' -line? | non_test | output a and e in tga analysis mode simulations with tga analysis true usually start with reference temperature etc first order but if user wants to see the resulting a and e and continue by specifying modifying them directly it would be convenient to see the calculated values from somewhere normally fds writes a and e in out file under material definitions but tga analysis mode does not produce out file finding a and e requires removing tga analysis true line and making sure the file runs otherwise unnecessary steps in the process how about writing a and e directly to lu err before the stop fds performed line | 0 |
192,259 | 14,612,792,712 | IssuesEvent | 2020-12-22 06:56:24 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | YanshuoH/youkonger-backend: dao/event_date_test.go; 4 LoC | fresh test tiny |
Found a possible issue in [YanshuoH/youkonger-backend](https://www.github.com/YanshuoH/youkonger-backend) at [dao/event_date_test.go](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 53 may store a reference to ep
[Click here to see the code in its original context.](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54)
<details>
<summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ep := range epToInserts {
ep.EventDateID = edToInsert.ID
Expect(Conn.Create(&ep).Error).ToNot(HaveOccurred())
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: dcd1175b9976855007d9c3cdc9c70ec685dd5f14
| 1.0 | YanshuoH/youkonger-backend: dao/event_date_test.go; 4 LoC -
Found a possible issue in [YanshuoH/youkonger-backend](https://www.github.com/YanshuoH/youkonger-backend) at [dao/event_date_test.go](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 53 may store a reference to ep
[Click here to see the code in its original context.](https://github.com/YanshuoH/youkonger-backend/blob/dcd1175b9976855007d9c3cdc9c70ec685dd5f14/dao/event_date_test.go#L51-L54)
<details>
<summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ep := range epToInserts {
ep.EventDateID = edToInsert.ID
Expect(Conn.Create(&ep).Error).ToNot(HaveOccurred())
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: dcd1175b9976855007d9c3cdc9c70ec685dd5f14
| test | yanshuoh youkonger backend dao event date test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call at line may store a reference to ep click here to show the line s of go which triggered the analyzer go for ep range eptoinserts ep eventdateid edtoinsert id expect conn create ep error tonot haveoccurred leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
179,189 | 13,850,131,129 | IssuesEvent | 2020-10-15 00:20:11 | xbmc/metadata.tvshows.themoviedb.org.python | https://api.github.com/repos/xbmc/metadata.tvshows.themoviedb.org.python | closed | Episode Titles not scraped with fallback language | bug testing | Issue reported here for the XML scraper also affects this Python scraper...
https://github.com/xbmc/repo-scrapers/issues/225 | 1.0 | Episode Titles not scraped with fallback language - Issue reported here for the XML scraper also affects this Python scraper...
https://github.com/xbmc/repo-scrapers/issues/225 | test | episode titles not scraped with fallback language issue reported here for the xml scraper also affects this python scraper | 1 |
182,379 | 14,117,051,253 | IssuesEvent | 2020-11-08 07:07:42 | fabioturazzidouglas/Hackermen_Project | https://api.github.com/repos/fabioturazzidouglas/Hackermen_Project | closed | JUnit Booking System Testing | Testing | Tests the iteration between all models (Users, Bookings, Rooms, RoomTypes) to create bookings | 1.0 | JUnit Booking System Testing - Tests the iteration between all models (Users, Bookings, Rooms, RoomTypes) to create bookings | test | junit booking system testing tests the iteration between all models users bookings rooms roomtypes to create bookings | 1 |
126,230 | 10,413,013,799 | IssuesEvent | 2019-09-13 17:26:35 | futest-test/fu | https://api.github.com/repos/futest-test/fu | opened | Vulnerability - Cross site scripting | FoundByAcunetix360 FuTest |
**URL:** http://php.testsparker.com/
**Name:** Cross site scripting
**Severity:** High
You can see vulnerability details from the link below.
http://ec2-18-194-173-226.eu-central-1.compute.amazonaws.com/vulnerabilities/detail/df6c902af25242c98a98aac602aa83ff | 1.0 | Vulnerability - Cross site scripting -
**URL:** http://php.testsparker.com/
**Name:** Cross site scripting
**Severity:** High
You can see vulnerability details from the link below.
http://ec2-18-194-173-226.eu-central-1.compute.amazonaws.com/vulnerabilities/detail/df6c902af25242c98a98aac602aa83ff | test | vulnerability cross site scripting url name cross site scripting severity high you can see vulnerability details from the link below | 1 |
548,545 | 16,066,511,475 | IssuesEvent | 2021-04-23 20:03:16 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Cannot Delete Last workflow survey question | component:ui priority:high state:in_progress type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
<img width="1483" alt="Screen Shot 2020-07-20 at 1 25 35 PM" src="https://user-images.githubusercontent.com/9889020/87967264-93ac5b80-ca8c-11ea-8050-66400bb6a019.png">
If you create a survey question for a workflow job template and then try to delete that question you get the above error. It appears as if we're trying to delete the question from `/api/v2/job_templates/:id/survey_spec` instead of `/api/v2/workflow_job_templates/:id/survey_spec`
##### STEPS TO REPRODUCE
1. Create a survey for a WFJT
2. Enable surveys for WFJT
3. Select Survey
4. Attempt to delete Survey
Note: 404
also
1. Create many surveys for a WFJT
2. Enable surveys for WFJT
3. Select All surveys
4. Attempt to delete all surveys
Note: 404 | 1.0 | Cannot Delete Last workflow survey question - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
<img width="1483" alt="Screen Shot 2020-07-20 at 1 25 35 PM" src="https://user-images.githubusercontent.com/9889020/87967264-93ac5b80-ca8c-11ea-8050-66400bb6a019.png">
If you create a survey question for a workflow job template and then try to delete that question you get the above error. It appears as if we're trying to delete the question from `/api/v2/job_templates/:id/survey_spec` instead of `/api/v2/workflow_job_templates/:id/survey_spec`
##### STEPS TO REPRODUCE
1. Create a survey for a WFJT
2. Enable surveys for WFJT
3. Select Survey
4. Attempt to delete Survey
Note: 404
also
1. Create many surveys for a WFJT
2. Enable surveys for WFJT
3. Select All surveys
4. Attempt to delete all surveys
Note: 404 | non_test | cannot delete last workflow survey question issue type bug report summary img width alt screen shot at pm src if you create a survey question for a workflow job template and then try to delete that question you get the above error it appears as if we re trying to delete the question from api job templates id survey spec instead of api workflow job templates id survey spec steps to reproduce create a survey for a wfjt enable surveys for wfjt select survey attempt to delete survey note also create many surveys for a wfjt enable surveys for wfjt select all surveys attempt to delete all surveys note | 0 |
151,315 | 12,031,686,994 | IssuesEvent | 2020-04-13 10:15:58 | microsoft/azure-tools-for-java | https://api.github.com/repos/microsoft/azure-tools-for-java | closed | Can't trigger function/admin http function when click 'Trigger Function' button | External Dependency Internal Test function | #### Environment:
- OS : Win&Mac
- Plugins Version : Azure java toolkit for IntelliJ: 3.34.0
- IDE Version: IU-jre8 & IC-jre11
#### Repro steps:
1. Create a function project, add function/admin auth type http trigger and run deploy.
2. Go to azure explorer, find the function app and right click 'Trigger Function' button.
#### Result:

| 1.0 | Can't trigger function/admin http function when click 'Trigger Function' button - #### Environment:
- OS : Win&Mac
- Plugins Version : Azure java toolkit for IntelliJ: 3.34.0
- IDE Version: IU-jre8 & IC-jre11
#### Repro steps:
1. Create a function project, add function/admin auth type http trigger and run deploy.
2. Go to azure explorer, find the function app and right click 'Trigger Function' button.
#### Result:

| test | can t trigger function admin http function when click trigger function button environment: os win mac plugins version azure java toolkit for intellij ide version iu ic repro steps create a function project add function admin auth type http trigger and run deploy go to azure explorer find the function app and right click trigger function button result | 1 |
110,357 | 13,906,716,633 | IssuesEvent | 2020-10-20 11:39:36 | httpwg/httpbis-issues | https://api.github.com/repos/httpwg/httpbis-issues | opened | Ordering in Upgrade (Trac #445) | Incomplete Migration Migrated from Trac design draft-ietf-httpbis-p1-messaging@tools.ietf.org p1-messaging | Migrated from https://trac.ietf.org/ticket/445
```json
{
"status": "closed",
"changetime": "2013-09-26T05:56:39",
"_ts": "1380174999465154",
"description": "p1 section 6.7 defines the Upgrade header, but no where does it say anything about relative preference.\n\nShould we define (or at least allow) for the ordering to be semantically significant? It seems to me that if we end up using this, and there are a few different variants of HTTP/2 (e.g., \"normal\" vs \"mobile\"), it'd be nice to rely on ordering here.",
"reporter": "mnot@pobox.com",
"cc": "",
"resolution": "fixed",
"time": "2013-04-20T07:51:45",
"component": "p1-messaging",
"summary": "Ordering in Upgrade",
"priority": "normal",
"keywords": "",
"milestone": "23",
"owner": "draft-ietf-httpbis-p1-messaging@tools.ietf.org",
"type": "design",
"severity": "In WG Last Call"
}
```
| 1.0 | Ordering in Upgrade (Trac #445) - Migrated from https://trac.ietf.org/ticket/445
```json
{
"status": "closed",
"changetime": "2013-09-26T05:56:39",
"_ts": "1380174999465154",
"description": "p1 section 6.7 defines the Upgrade header, but no where does it say anything about relative preference.\n\nShould we define (or at least allow) for the ordering to be semantically significant? It seems to me that if we end up using this, and there are a few different variants of HTTP/2 (e.g., \"normal\" vs \"mobile\"), it'd be nice to rely on ordering here.",
"reporter": "mnot@pobox.com",
"cc": "",
"resolution": "fixed",
"time": "2013-04-20T07:51:45",
"component": "p1-messaging",
"summary": "Ordering in Upgrade",
"priority": "normal",
"keywords": "",
"milestone": "23",
"owner": "draft-ietf-httpbis-p1-messaging@tools.ietf.org",
"type": "design",
"severity": "In WG Last Call"
}
```
| non_test | ordering in upgrade trac migrated from json status closed changetime ts description section defines the upgrade header but no where does it say anything about relative preference n nshould we define or at least allow for the ordering to be semantically significant it seems to me that if we end up using this and there are a few different variants of http e g normal vs mobile it d be nice to rely on ordering here reporter mnot pobox com cc resolution fixed time component messaging summary ordering in upgrade priority normal keywords milestone owner draft ietf httpbis messaging tools ietf org type design severity in wg last call | 0 |
38,548 | 6,676,677,711 | IssuesEvent | 2017-10-05 07:14:48 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Python3 doctest compatibility: algorithms IntegrateByComponent to Integration | Component: Documentation Misc: Maintenance | This issue is part of the move to Python3 in the doctests,
Algorithms covered in this issue are:
- [x] IntegrateByComponent
- [x] IntegrateEPP
- [x] IntegrateEllipsoids
- [x] IntegrateEllipsoidsTwoStep
- [x] IntegrateFlux
- [x] IntegrateMDHistoWorkspace
- [x] IntegratePeakTimeSlices
- [x] IntegratePeaksCWSD
- [x] IntegratePeaksHybrid
- [x] IntegratePeaksMD
- [x] IntegratePeaksMDHKL
- [x] IntegratePeaksUsingClusters
- [x] Integration
| 1.0 | Python3 doctest compatibility: algorithms IntegrateByComponent to Integration - This issue is part of the move to Python3 in the doctests,
Algorithms covered in this issue are:
- [x] IntegrateByComponent
- [x] IntegrateEPP
- [x] IntegrateEllipsoids
- [x] IntegrateEllipsoidsTwoStep
- [x] IntegrateFlux
- [x] IntegrateMDHistoWorkspace
- [x] IntegratePeakTimeSlices
- [x] IntegratePeaksCWSD
- [x] IntegratePeaksHybrid
- [x] IntegratePeaksMD
- [x] IntegratePeaksMDHKL
- [x] IntegratePeaksUsingClusters
- [x] Integration
| non_test | doctest compatibility algorithms integratebycomponent to integration this issue is part of the move to in the doctests algorithms covered in this issue are integratebycomponent integrateepp integrateellipsoids integrateellipsoidstwostep integrateflux integratemdhistoworkspace integratepeaktimeslices integratepeakscwsd integratepeakshybrid integratepeaksmd integratepeaksmdhkl integratepeaksusingclusters integration | 0 |
318,684 | 27,321,017,404 | IssuesEvent | 2023-02-24 19:50:52 | peviitor-ro/ui-js | https://api.github.com/repos/peviitor-ro/ui-js | closed | [SERP] Different font family for the text "Alătură-te" | bug TestQuality | ## Precondition
URL: https://beta.peviitor.ro/
Device: Samsung Galaxy S21 Ultra
Browser: Chrome
Platform: Android 12
## Steps to Reproduce:
### Step 1 <span style="color:#58b880"> **[Pass]** </span>
Open URL in browser
#### Expected Result
Website is loaded without any error
### Step 2 <span style="color:#58b880"> **[Pass]** </span>
Click on “Caută”
#### Expected Result
The user is redirected to SERP
### Step 3 <span style="color:#ff5538"> **[Fail]** </span>
Inspect the font of the text "Alătură-te"
#### Expected Result
Font is 'SF Pro'
#### Actual Result
The actual font is SegoeUI
| 1.0 | [SERP] Different font family for the text "Alătură-te" - ## Precondition
URL: https://beta.peviitor.ro/
Device: Samsung Galaxy S21 Ultra
Browser: Chrome
Platform: Android 12
## Steps to Reproduce:
### Step 1 <span style="color:#58b880"> **[Pass]** </span>
Open URL in browser
#### Expected Result
Website is loaded without any error
### Step 2 <span style="color:#58b880"> **[Pass]** </span>
Click on “Caută”
#### Expected Result
The user is redirected to SERP
### Step 3 <span style="color:#ff5538"> **[Fail]** </span>
Inspect the font of the text "Alătură-te"
#### Expected Result
Font is 'SF Pro'
#### Actual Result
The actual font is SegoeUI
| test | different font family for the text alătură te precondition url device samsung galaxy ultra browser chrome platform android steps to reproduce step open url in browser expected result website is loaded without any error step click on “caută” expected result the user is redirected to serp step inspect the font of the text quot alătură te quot expected result font is sf pro actual result the actual font is segoeui | 1 |
118,029 | 15,216,440,567 | IssuesEvent | 2021-02-17 15:30:39 | wordpress-mobile/WordPress-Android | https://api.github.com/repos/wordpress-mobile/WordPress-Android | opened | Update Activity Log Details UI and Add Backup Download Button/Flow | Jetpack Jetpack Mobile Needs Design [Type] Enhancement | As per the requirements in this post (`pcdRpT-cM`, see `Activity Log: Event Details`), this issue is about:
- UI updates on this screen to match design, and
- Adding the `Backup Download` button and connecting the button with the new flow.
Before | After
-------|------
<img width="250" height="500" alt="before" src="https://user-images.githubusercontent.com/9729923/108225084-4092a800-7144-11eb-9458-e2ece9899626.png"> | <img width="250" height="500" alt="after" src="https://user-images.githubusercontent.com/9729923/108225150-53a57800-7144-11eb-88ef-dce217851bd4.png">
On Android, the screen looks a bit different today, comparing the before (or current) of Android with the after (or current) of iOS:
Before or Current (Android) | After or Current (iOS)
-------|------
<img width="250" height="500" alt="before android" src="https://user-images.githubusercontent.com/9729923/108225656-cf9fc000-7144-11eb-9232-0fc8cb8126a1.png"> | <img width="250" height="500" alt="after ios" src="https://user-images.githubusercontent.com/9729923/108226033-302efd00-7145-11eb-8eaf-883176d46547.png">
@osullivanchris can you please suggest what UI changes need to happen to the UI so that Android get the appropriate after look, and still be comparable with iOS's after look?
| 1.0 | Update Activity Log Details UI and Add Backup Download Button/Flow - As per the requirements in this post (`pcdRpT-cM`, see `Activity Log: Event Details`), this issue is about:
- UI updates on this screen to match design, and
- Adding the `Backup Download` button and connecting the button with the new flow.
Before | After
-------|------
<img width="250" height="500" alt="before" src="https://user-images.githubusercontent.com/9729923/108225084-4092a800-7144-11eb-9458-e2ece9899626.png"> | <img width="250" height="500" alt="after" src="https://user-images.githubusercontent.com/9729923/108225150-53a57800-7144-11eb-88ef-dce217851bd4.png">
On Android, the screen looks a bit different today, comparing the before (or current) of Android with the after (or current) of iOS:
Before or Current (Android) | After or Current (iOS)
-------|------
<img width="250" height="500" alt="before android" src="https://user-images.githubusercontent.com/9729923/108225656-cf9fc000-7144-11eb-9232-0fc8cb8126a1.png"> | <img width="250" height="500" alt="after ios" src="https://user-images.githubusercontent.com/9729923/108226033-302efd00-7145-11eb-8eaf-883176d46547.png">
@osullivanchris can you please suggest what UI changes need to happen to the UI so that Android get the appropriate after look, and still be comparable with iOS's after look?
| non_test | update activity log details ui and add backup download button flow as per the requirements in this post pcdrpt cm see activity log event details this issue is about ui updates on this screen to match design and adding the backup download button and connecting the button with the new flow before after img width height alt before src img width height alt after src on android the screen looks a bit different today comparing the before or current of android with the after or current of ios before or current android after or current ios img width height alt before android src img width height alt after ios src osullivanchris can you please suggest what ui changes need to happen to the ui so that android get the appropriate after look and still be comparable with ios s after look | 0 |
4,416 | 3,369,801,513 | IssuesEvent | 2015-11-23 12:10:36 | mozilla/feedthefox | https://api.github.com/repos/mozilla/feedthefox | closed | Show "code name" on Devices pages | Builds/Devices Discussion UI/UX | It might be good idea to tell people the "code name" of their device at https://firefoxos.mozilla.community/devices/ because they're most likely going to need that bit of information once they click the big green buttons.
The ones I know of so far FWIW:
Xperia Z3 (leo)
Xperia Z3 Compact (aries)
Xperia Z2 (sirius)
Xperia T2 Ultra (tianchi)
Xperia E3 (flamingo) | 1.0 | Show "code name" on Devices pages - It might be good idea to tell people the "code name" of their device at https://firefoxos.mozilla.community/devices/ because they're most likely going to need that bit of information once they click the big green buttons.
The ones I know of so far FWIW:
Xperia Z3 (leo)
Xperia Z3 Compact (aries)
Xperia Z2 (sirius)
Xperia T2 Ultra (tianchi)
Xperia E3 (flamingo) | non_test | show code name on devices pages it might be good idea to tell people the code name of their device at because they re most likely going to need that bit of information once they click the big green buttons the ones i know of so far fwiw xperia leo xperia compact aries xperia sirius xperia ultra tianchi xperia flamingo | 0 |
28,115 | 4,365,170,922 | IssuesEvent | 2016-08-03 09:48:38 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | circleci: failed tests: TestDockerPython | Robot test-failure | The following test appears to have failed:
[#20875](https://circleci.com/gh/cockroachdb/cockroach/20875):
```
^
SyntaxError: invalid syntax
I160802 15:54:31.215544 acceptance/cluster/localcluster.go:673 stopping
I160802 15:54:31.215794 acceptance/cluster/localcluster.go:573 event stream done, resetting...: net/http: request canceled
I160802 15:54:31.215807 acceptance/cluster/localcluster.go:596 events monitor exits
--- FAIL: TestDockerPython (20.37s)
panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist [recovered]
panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist
goroutine 2703 [running]:
panic(0x14919a0, 0xc8204a3210)
/usr/local/go/src/runtime/panic.go:481 +0x3e6
testing.tRunner.func1(0xc8202b0360)
/usr/local/go/src/testing/testing.go:467 +0x192
panic(0x14919a0, 0xc8204a3210)
/usr/local/go/src/runtime/panic.go:443 +0x4e9
github.com/cockroachdb/cockroach/acceptance/cluster.maybePanic(0x7fca667c1028, 0xc8204a3210)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/docker.go:166 +0x4b
github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).stop(0xc8204641c0)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:699 +0x620
github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).AssertAndStop(0xc8204641c0, 0xc8202b0360)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:664 +0x53
github.com/cockroachdb/cockroach/acceptance.testDocker(0xc8202b0360, 0x1, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x7fca66789408, 0xc82042eab0)
/go/src/github.com/cockroachdb/cockroach/acceptance/util_test.go:406 +0x5bc
github.com/cockroachdb/cockroach/acceptance.testDockerSingleNode(0xc8202b0360, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x0, 0x0)
```
Please assign, take a look and update the issue accordingly. | 1.0 | circleci: failed tests: TestDockerPython - The following test appears to have failed:
[#20875](https://circleci.com/gh/cockroachdb/cockroach/20875):
```
^
SyntaxError: invalid syntax
I160802 15:54:31.215544 acceptance/cluster/localcluster.go:673 stopping
I160802 15:54:31.215794 acceptance/cluster/localcluster.go:573 event stream done, resetting...: net/http: request canceled
I160802 15:54:31.215807 acceptance/cluster/localcluster.go:596 events monitor exits
--- FAIL: TestDockerPython (20.37s)
panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist [recovered]
panic: Error response from daemon: Cannot kill container a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c: active container for a767718f508ad84f273a08c48ab1325e7c04df5528b08f2744c29cf34291275c does not exist
goroutine 2703 [running]:
panic(0x14919a0, 0xc8204a3210)
/usr/local/go/src/runtime/panic.go:481 +0x3e6
testing.tRunner.func1(0xc8202b0360)
/usr/local/go/src/testing/testing.go:467 +0x192
panic(0x14919a0, 0xc8204a3210)
/usr/local/go/src/runtime/panic.go:443 +0x4e9
github.com/cockroachdb/cockroach/acceptance/cluster.maybePanic(0x7fca667c1028, 0xc8204a3210)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/docker.go:166 +0x4b
github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).stop(0xc8204641c0)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:699 +0x620
github.com/cockroachdb/cockroach/acceptance/cluster.(*LocalCluster).AssertAndStop(0xc8204641c0, 0xc8202b0360)
/go/src/github.com/cockroachdb/cockroach/acceptance/cluster/localcluster.go:664 +0x53
github.com/cockroachdb/cockroach/acceptance.testDocker(0xc8202b0360, 0x1, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x7fca66789408, 0xc82042eab0)
/go/src/github.com/cockroachdb/cockroach/acceptance/util_test.go:406 +0x5bc
github.com/cockroachdb/cockroach/acceptance.testDockerSingleNode(0xc8202b0360, 0x1661060, 0x6, 0xc82042e180, 0x3, 0x3, 0x0, 0x0)
```
Please assign, take a look and update the issue accordingly. | test | circleci failed tests testdockerpython the following test appears to have failed syntaxerror invalid syntax acceptance cluster localcluster go stopping acceptance cluster localcluster go event stream done resetting net http request canceled acceptance cluster localcluster go events monitor exits fail testdockerpython panic error response from daemon cannot kill container active container for does not exist panic error response from daemon cannot kill container active container for does not exist goroutine panic usr local go src runtime panic go testing trunner usr local go src testing testing go panic usr local go src runtime panic go github com cockroachdb cockroach acceptance cluster maybepanic go src github com cockroachdb cockroach acceptance cluster docker go github com cockroachdb cockroach acceptance cluster localcluster stop go src github com cockroachdb cockroach acceptance cluster localcluster go github com cockroachdb cockroach acceptance cluster localcluster assertandstop go src github com cockroachdb cockroach acceptance cluster localcluster go github com cockroachdb cockroach acceptance testdocker go src github com cockroachdb cockroach acceptance util test go github com cockroachdb cockroach acceptance testdockersinglenode please assign take a look and update the issue accordingly | 1 |
322,301 | 27,595,465,101 | IssuesEvent | 2023-03-09 05:50:46 | antrea-io/antrea | https://api.github.com/repos/antrea-io/antrea | opened | windows e2e test used stale image on one Node | area/test area/test/e2e area/OS/windows kind/bug | **Describe the bug**
<!--
A clear and concise description of what the bug is.
If you believe this bug is a security issue, please don't use this template and follow our [security guidelines](/SECURITY.md)
-->
When I was debugging an issue on a windows testbed, I found one antrea-agent-windows Pod ran an old image. The correct version should be v1.11.0-dev-75a53913.dirty:
```
# antrea-agent-windows-nrpvd
I0308 21:10:38.550971 6844 agent.go:99] Starting Antrea agent (version v1.11.0-dev-75a53913.dirty)
# antrea-agent-windows-ztlsr
I0308 21:09:47.755164 1484 agent.go:92] Starting Antrea agent (version v1.7.0-dev-2a37aec1.dirty)
```
Testbed deployment log:
```
[antrea-windows-e2e-proxyall-for-pull-request] $ /bin/sh -xe /tmp/jenkins8184209255165314312.sh
+ set -e
+ head -n1 ci/docker-registry
+ DOCKER_REGISTRY=projects.registry.vmware.com
+ sed -i /skipIfNotIPv4Cluster(t)/a\ \ \ \ \ \ \ \ skipIfHasWindowsNodes(t) test/e2e/nodeportlocal_test.go
+ ./ci/jenkins/test.sh --testcase windows-e2e --registry projects.registry.vmware.com --proxyall
===== Clean up stale files & folders older than 7 days under /tmp =====
====== Cleanup Antrea Installation ======
No resources found in antrea-test namespace.
====== Building Antrea for the Following Commit ======
commit 75a5391315ac5d89f2a1e57f3cecbca3f816ade5
Merge: 32485497 18ddb089
Author: Kumar Atish <atish.iaf@gmail.com>
Date: Wed Mar 8 23:03:12 2023 +0530
Merge 18ddb08974afc6d051c489fb87890a495841cb56 into 324854979d383c11e83e49b5c7e88d1caec6ebdd
7 1 pkg/agent/apiserver/handlers/memberlist/handler.go
62 26 pkg/agent/apiserver/handlers/memberlist/handler_test.go
5 1 pkg/support/dump.go
Deleted Images:
untagged: antrea/cni-binaries:v1.1.1
untagged: projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
untagged: projects.registry.vmware.com/antrea/cni-binaries@sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219
deleted: sha256:bfaa431b732dd3d12776c98d0e684f65e9cf77c481b9e35496b3e1b02e25da05
deleted: sha256:fe0b72db62131760671074f192f8c182a0c22c6fa2292316123ebdf71ddb77f5
deleted: sha256:55139abd5e2a8db4f73283a7f50ed7b3b1d81d4d2956ec119a3d1f03c2eaeb4d
deleted: sha256:7e7fc777a83c92df7fb369f3b2a61189218f567944bbf9269f69453372c109d9
deleted: sha256:25bb6906eea3f717d8295b99a4d69538624cedafa62c04bf7dff0b4ed97214d3
deleted: sha256:f6d6f546f40138b7b433a17e74af250fa533cbd34e24783d0ceef0a1b7fc7b64
deleted: sha256:67ce281807736e142e8151f56dd4a2e387d9389449c170eb7405c04761479ed4
untagged: ubuntu:22.04
untagged: projects.registry.vmware.com/antrea/ubuntu:22.04
untagged: projects.registry.vmware.com/antrea/ubuntu@sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b
untagged: e2eteam/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1
deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4
deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9
deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3
deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88
deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac
deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf
deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463
deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68
deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc
deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac
untagged: antrea/openvswitch-debs:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch-debs@sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668
deleted: sha256:fd116dd5d80d4f166894a08e1832afaf52fbbd6367dad185199094c996f62301
deleted: sha256:81f789fce63bda6d07e049856d386c61a636528699c76d022708ddee445b4b2e
deleted: sha256:6b8158b5883be250d6f351957fb43c33a4e4c4b63898aa7a2a86522e0f0dc523
deleted: sha256:2eb4d98e07106613c35d3a505adcf490871ceeb3de9ce520c3b5ed1ab757ebf1
deleted: sha256:c960d4cb5b1e6437b012bf69ee311156c1b63ce80dae9bec4277b26b9947b742
deleted: sha256:e2b36c6bf533a216accfb28a94c4acb21ee631cab9b8eab1d723e65ce176d71a
deleted: sha256:a68b22d0c835019f3f34a2eeabf873611708513b616483c091d94673ed9cc329
untagged: nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101
deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478
deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f
deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a
deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282
deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81
untagged: antrea/openvswitch:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch@sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83
untagged: antrea/base-ubuntu:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/base-ubuntu@sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d
Total reclaimed space: 773.4MB
BUILD_TAG: antrea-v1.11
22.04: Pulling from antrea/ubuntu
Digest: sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b
Status: Downloaded newer image for projects.registry.vmware.com/antrea/ubuntu:22.04
projects.registry.vmware.com/antrea/ubuntu:22.04
1.19: Pulling from antrea/golang
Digest: sha256:6687844637736cc26e4a10386abb7b482e5d231574be0bef9260a255da1f4775
Status: Image is up to date for projects.registry.vmware.com/antrea/golang:1.19
projects.registry.vmware.com/antrea/golang:1.19
antrea-v1.11: Pulling from antrea/openvswitch-debs
76769433fd8a: Already exists
57d34a411582: Pulling fs layer
7dad3cb1dcc7: Pulling fs layer
4f21db537580: Pulling fs layer
7dad3cb1dcc7: Verifying Checksum
7dad3cb1dcc7: Download complete
4f21db537580: Verifying Checksum
4f21db537580: Download complete
57d34a411582: Verifying Checksum
57d34a411582: Download complete
57d34a411582: Pull complete
7dad3cb1dcc7: Pull complete
4f21db537580: Pull complete
Digest: sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668
Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
antrea-v1.11: Pulling from antrea/openvswitch
Digest: sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83
Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
v1.1.1: Pulling from antrea/cni-binaries
76769433fd8a: Already exists
c96ff6af3775: Pulling fs layer
247dd6786f33: Pulling fs layer
247dd6786f33: Verifying Checksum
247dd6786f33: Download complete
c96ff6af3775: Verifying Checksum
c96ff6af3775: Download complete
c96ff6af3775: Pull complete
247dd6786f33: Pull complete
Digest: sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219
Status: Downloaded newer image for projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
antrea-v1.11: Pulling from antrea/base-ubuntu
Digest: sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d
Status: Downloaded newer image for projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
Sending build context to Docker daemon 24.06kB
Step 1/5 : FROM ubuntu:22.04 as ovs-debs
---> 74f2314a03de
Step 2/5 : ARG OVS_VERSION
---> Using cache
---> a68b22d0c835
Step 3/5 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev
---> Using cache
---> c960d4cb5b1e
Step 4/5 : COPY apply-patches.sh /
---> Using cache
---> 6b8158b5883b
Step 5/5 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch*
---> Using cache
---> fd116dd5d80d
Successfully built fd116dd5d80d
Successfully tagged antrea/openvswitch-debs:antrea-v1.11
Sending build context to Docker daemon 24.06kB
Step 1/11 : FROM ubuntu:22.04 as ovs-debs
---> 74f2314a03de
Step 2/11 : ARG OVS_VERSION
---> Using cache
---> a68b22d0c835
Step 3/11 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev
---> Using cache
---> c960d4cb5b1e
Step 4/11 : COPY apply-patches.sh /
---> Using cache
---> 6b8158b5883b
Step 5/11 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch*
---> Using cache
---> fd116dd5d80d
Step 6/11 : FROM ubuntu:22.04
---> 74f2314a03de
Step 7/11 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> 2873291cca14
Step 8/11 : LABEL description="A Docker image based on Ubuntu 22.04 which includes Open vSwitch built from source."
---> Using cache
---> 0e26ec9b7d9d
Step 9/11 : COPY --from=ovs-debs /tmp/ovs-debs/* /tmp/ovs-debs/
---> Using cache
---> 0c4db4416cce
Step 10/11 : COPY charon-logging.conf /tmp
---> Using cache
---> 79d23217a023
Step 11/11 : RUN apt-get update && apt-get install -y --no-install-recommends iptables logrotate libstrongswan-standard-plugins && (dpkg -i /tmp/ovs-debs/*.deb || apt-get -f -y --no-install-recommends install) && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && sed -i "/rotate /a\ #size 100M" /etc/logrotate.d/openvswitch-switch && sed -i "/^.*filelog.*{/r /tmp/charon-logging.conf" /etc/strongswan.d/charon-logging.conf && rm -rf /tmp/*
---> Using cache
---> 3852ac294105
Successfully built 3852ac294105
Successfully tagged antrea/openvswitch:antrea-v1.11
/var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request
Sending build context to Docker daemon 13.31kB
Step 1/7 : ARG BUILD_TAG
Step 2/7 : FROM ubuntu:22.04 as cni-binaries
---> 74f2314a03de
Step 3/7 : ARG CNI_BINARIES_VERSION
---> Using cache
---> 67ce28180773
Step 4/7 : ARG WHEREABOUTS_VERSION=v0.5.4
---> Using cache
---> f6d6f546f401
Step 5/7 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates
---> Using cache
---> 7e7fc777a83c
Step 6/7 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth"
---> Using cache
---> 55139abd5e2a
Step 7/7 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts
---> Using cache
---> bfaa431b732d
Successfully built bfaa431b732d
Successfully tagged antrea/cni-binaries:v1.1.1
Sending build context to Docker daemon 13.31kB
Step 1/15 : ARG BUILD_TAG
Step 2/15 : FROM ubuntu:22.04 as cni-binaries
---> 74f2314a03de
Step 3/15 : ARG CNI_BINARIES_VERSION
---> Using cache
---> 67ce28180773
Step 4/15 : ARG WHEREABOUTS_VERSION=v0.5.4
---> Using cache
---> f6d6f546f401
Step 5/15 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates
---> Using cache
---> 7e7fc777a83c
Step 6/15 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth"
---> Using cache
---> 55139abd5e2a
Step 7/15 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts
---> Using cache
---> bfaa431b732d
Step 8/15 : FROM antrea/openvswitch:${BUILD_TAG}
---> 3852ac294105
Step 9/15 : ARG SURICATA_VERSION
---> Using cache
---> bafddc6af04e
Step 10/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> ec89fc76d92d
Step 11/15 : LABEL description="An Ubuntu based Docker base image for Antrea."
---> Using cache
---> dd8fb640925d
Step 12/15 : USER root
---> Using cache
---> 5c814d44d69b
Step 13/15 : ADD https://raw.githubusercontent.com/kubernetes-sigs/iptables-wrappers/9e6ce59c864623ea71a6f7d59c35fcb13a919b87/iptables-wrapper-installer.sh /iptables-wrapper-installer.sh
---> Using cache
---> 789847397959
Step 14/15 : RUN apt-get update && apt-get install -y --no-install-recommends ipset jq inotify-tools gpg-agent software-properties-common && add-apt-repository ppa:oisf/suricata-${SURICATA_VERSION} && apt-get update && apt-get install -y suricata && apt-get remove -y gpg-agent software-properties-common && apt-get autoremove -y && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && chmod +x /iptables-wrapper-installer.sh && /iptables-wrapper-installer.sh
---> Using cache
---> 815257da934e
Step 15/15 : COPY --from=cni-binaries /opt/cni/bin /opt/cni/bin
---> Using cache
---> ad75b5563247
Successfully built ad75b5563247
Successfully tagged antrea/base-ubuntu:antrea-v1.11
/var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request
===> Building Antrea bins and antrea/antrea-ubuntu Docker image <===
docker build -t antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.ubuntu --build-arg OVS_VERSION=2.17.3 --build-arg GO_VERSION=1.19 --build-arg BUILD_TAG=antrea-v1.11 .
Sending build context to Docker daemon 47.41MB
Step 1/15 : ARG GO_VERSION
Step 2/15 : ARG BUILD_TAG
Step 3/15 : FROM golang:${GO_VERSION} as antrea-build
---> ff3cd58379d9
Step 4/15 : WORKDIR /antrea
---> Running in 62533a8c8585
Removing intermediate container 62533a8c8585
---> 869906fadac2
Step 5/15 : COPY go.mod /antrea/go.mod
---> e087b2973e1c
Step 6/15 : RUN go mod download
---> Running in 6771daf43c3b
Removing intermediate container 6771daf43c3b
---> 41125c6f24ee
Step 7/15 : COPY . /antrea
---> de288ca18a4c
Step 8/15 : RUN make antrea-agent antrea-controller antrea-cni antctl-linux
---> Running in a3938237789d
GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-agent
GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-controller
GOOS=linux CGO_ENABLED=0 go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni
Removing intermediate container a3938237789d
---> 848dec0d7aaf
Step 9/15 : RUN mv bin/antctl-linux bin/antctl
---> Running in b5313b457af3
Removing intermediate container b5313b457af3
---> f6bfc1318921
Step 10/15 : FROM antrea/base-ubuntu:${BUILD_TAG}
---> ad75b5563247
Step 11/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> 54aa6d3d57c4
Step 12/15 : LABEL description="The Docker image to deploy the Antrea CNI."
---> Using cache
---> 8161ff16a1cb
Step 13/15 : USER root
---> Using cache
---> ef8b9278df6d
Step 14/15 : COPY build/images/scripts/* /usr/local/bin/
---> Using cache
---> afd46dfce8c0
Step 15/15 : COPY --from=antrea-build /antrea/bin/* /usr/local/bin/
---> 2d3d7425fc7f
[Warning] One or more build-args [OVS_VERSION] were not consumed
Successfully built 2d3d7425fc7f
Successfully tagged antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty
docker tag antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty antrea/antrea-ubuntu
====== Delivering Antrea to all Nodes ======
====== Updating yaml files to enable proxyAll ======
===== Pull necessary images on Control-Plane node =====
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/nginx:1.15-alpine
===== Deliver Antrea to Linux worker nodes and pull necessary images on worker nodes =====
sending incremental file list
antrea-ubuntu.tar
45,216 0% 0.00kB/s 0:00:00
158,854,896 31% 151.45MB/s 0:00:02
212,998,875 41% 101.60MB/s 0:00:02
282,112,496 55% 89.73MB/s 0:00:02
349,690,560 68% 83.42MB/s 0:00:01
402,000,296 78% 58.01MB/s 0:00:01
461,453,972 90% 59.28MB/s 0:00:00
511,165,952 100% 72.18MB/s 0:00:06 (xfr#1, to-chk=0/1)
sent 287,388,834 bytes received 158,320 bytes 33,829,076.94 bytes/sec
total size is 511,165,952 speedup is 1.78
Deleted Images:
untagged: k8s.gcr.io/e2e-test-images/agnhost:2.29
untagged: k8s.gcr.io/e2e-test-images/agnhost@sha256:1ae25df1a1985fa815f712b051f467dcd2d016d9efb74ca41ef385f9ed64ada9
deleted: sha256:3f4d3ab351ab4180a2c0fc655ba97263ba74fa353d41642b3e09bc16c231b735
deleted: sha256:9735e33cc1701e722d4ef42086090078232f4d45087ebea8a40de28b18eb190d
deleted: sha256:10f54c8640b3c21f6a7575ed67fa29c1811ed0a1404b0d398ab197ef2a55c46b
deleted: sha256:4104722c5b5696347fb190e2e341ac593a0dd3953ad63bf82c89e037f71da341
deleted: sha256:7abc3633883948ff206528cd56bada29a18222fd642bee4c76d67ab12e590113
deleted: sha256:41758cdb2ea620f0c2db17598ce779841c9d4dce6cb6c457b302e082b2d7771f
deleted: sha256:60e550c19b857d2261358edfa8b11cc43e53a076f3baea759553ce914889bd0d
deleted: sha256:aa75a9d85b48eba258a28d6975fb3db0393809cc290d3dab4d0bf2bdb97d1e79
deleted: sha256:891c281852b0ab60d0a307ea6c5fbd0697004ade3745e7adcfdd549391dd633f
deleted: sha256:33e8713114f88c8cb3f60c8a0a4aefe2500823b2fbbae05488b0185ba226caae
untagged: e2eteam/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1
deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4
deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9
deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3
deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88
deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac
deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf
deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463
deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68
deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc
deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac
untagged: nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101
deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478
deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f
deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a
deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282
deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81
Total reclaimed space: 253.1MB
The image antrea/antrea-ubuntu:latest already exists, renaming the old one with ID sha256:568e9f48739ea51ab066586797f29459d1710bf9608f8ab7acf591d99b77bf73 to empty string
Loaded image: antrea/antrea-ubuntu:latest
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/nginx:1.15-alpine
k8s.gcr.io/e2e-test-images/agnhost:2.29
===== Deliver Antrea Windows to Windows worker nodes and pull necessary images on Windows worker nodes =====
==== Reverting Windows VM a-ms-0008-win-0 =====
Windows VM a-ms-0008-win-0 powered on
projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.29
projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0
projects.registry.vmware.com/antrea/e2eteam-pause:3.2
Error response from daemon: Get https://mcr.microsoft.com/v2/: x509: certificate has expired or is not yet valid
=== Build Windows on Windows Node===
1.19-nanoserver: Pulling from antrea/golang
5c9d6483dab1: Pulling fs layer
9982991b8208: Pulling fs layer
cb374f741579: Pulling fs layer
732dbfa4c42c: Pulling fs layer
64bbe06a1892: Pulling fs layer
15263a6e39d7: Pulling fs layer
d1ebeb495bd3: Pulling fs layer
460984e65c46: Pulling fs layer
109a63e20d99: Pulling fs layer
e3c52c6cbb5c: Pulling fs layer
64bbe06a1892: Waiting
15263a6e39d7: Waiting
d1ebeb495bd3: Waiting
460984e65c46: Waiting
109a63e20d99: Waiting
e3c52c6cbb5c: Waiting
732dbfa4c42c: Waiting
9982991b8208: Verifying Checksum
9982991b8208: Download complete
cb374f741579: Verifying Checksum
cb374f741579: Download complete
732dbfa4c42c: Verifying Checksum
732dbfa4c42c: Download complete
64bbe06a1892: Verifying Checksum
64bbe06a1892: Download complete
15263a6e39d7: Verifying Checksum
15263a6e39d7: Download complete
d1ebeb495bd3: Verifying Checksum
d1ebeb495bd3: Download complete
109a63e20d99: Download complete
e3c52c6cbb5c: Verifying Checksum
e3c52c6cbb5c: Download complete
5c9d6483dab1: Verifying Checksum
5c9d6483dab1: Download complete
460984e65c46: Verifying Checksum
460984e65c46: Download complete
5c9d6483dab1: Pull complete
9982991b8208: Pull complete
cb374f741579: Pull complete
732dbfa4c42c: Pull complete
64bbe06a1892: Pull complete
15263a6e39d7: Pull complete
d1ebeb495bd3: Pull complete
460984e65c46: Pull complete
109a63e20d99: Pull complete
e3c52c6cbb5c: Pull complete
Digest: sha256:d5843e74790da22bd3d84d098dcbe21c7318a7c7c4db6bf1de09cb193e64c513
Status: Downloaded newer image for projects.registry.vmware.com/antrea/golang:1.19-nanoserver
projects.registry.vmware.com/antrea/golang:1.19-nanoserver
===> Building Antrea bins and antrea/antrea-windows Docker image <===
docker build --pull -t antrea/antrea-windows:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.windows --network host --build-arg GO_VERSION=1.19 --build-arg CNI_BINARIES_VERSION=v1.1.1 --build-arg NANOSERVER_VERSION=1809 --build-arg WIN_BUILD_TAG=caa8cac3d3 .
Sending build context to Docker daemon 558.6MB
Step 1/21 : ARG WIN_BUILD_TAG
Step 2/21 : ARG NANOSERVER_VERSION
Step 3/21 : FROM antrea/base-windows:${WIN_BUILD_TAG} as antrea-build-windows
caa8cac3d3: Pulling from antrea/base-windows
5ead999142ec: Pulling fs layer
3724c0f91ae7: Pulling fs layer
97a8c2800739: Pulling fs layer
c038a2fdb0b5: Pulling fs layer
5ef04caec17a: Pulling fs layer
8bca1755b69d: Pulling fs layer
3ebc2faa6c3e: Pulling fs layer
d85edb0b9152: Pulling fs layer
b343c39a6010: Pulling fs layer
b673c002e661: Pulling fs layer
99b72f4231b5: Pulling fs layer
d85edb0b9152: Waiting
3ebc2faa6c3e: Waiting
8bca1755b69d: Waiting
c038a2fdb0b5: Waiting
b343c39a6010: Waiting
5ef04caec17a: Waiting
b673c002e661: Waiting
99b72f4231b5: Waiting
97a8c2800739: Download complete
3724c0f91ae7: Download complete
5ead999142ec: Verifying Checksum
5ead999142ec: Download complete
c038a2fdb0b5: Download complete
5ef04caec17a: Verifying Checksum
5ef04caec17a: Download complete
8bca1755b69d: Verifying Checksum
8bca1755b69d: Download complete
b343c39a6010: Verifying Checksum
b343c39a6010: Download complete
3ebc2faa6c3e: Verifying Checksum
3ebc2faa6c3e: Download complete
99b72f4231b5: Verifying Checksum
99b72f4231b5: Download complete
b673c002e661: Verifying Checksum
b673c002e661: Download complete
d85edb0b9152: Verifying Checksum
d85edb0b9152: Download complete
5ead999142ec: Pull complete
3724c0f91ae7: Pull complete
97a8c2800739: Pull complete
c038a2fdb0b5: Pull complete
5ef04caec17a: Pull complete
8bca1755b69d: Pull complete
3ebc2faa6c3e: Pull complete
d85edb0b9152: Pull complete
b343c39a6010: Pull complete
b673c002e661: Pull complete
99b72f4231b5: Pull complete
Digest: sha256:c6edebf1a80553a5d474b6663657e69bc5b2cf54b116a1759e4d053c609a575d
Status: Downloaded newer image for antrea/base-windows:caa8cac3d3
---> 79d3453071f3
Step 4/21 : WORKDIR /antrea
---> Running in 8019bfcc22c9
Removing intermediate container 8019bfcc22c9
---> 6f713bbf1027
Step 5/21 : COPY go.mod /antrea/go.mod
---> 30bbece0963c
Step 6/21 : RUN go mod download
---> Running in 7a09a9c079f4
Removing intermediate container 7a09a9c079f4
---> af9b4babd43e
Step 7/21 : COPY . /antrea
---> 60d82c848851
Step 8/21 : RUN sh -c 'make windows-bin'
---> Running in aab89dd36ca9
GOOS=windows CGO_ENABLED=0 go build -o C:/antrea/bin -ldflags ' -X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni antrea.io/antrea/cmd/antrea-agent antrea.io/antrea/cmd/antctl
Removing intermediate container aab89dd36ca9
---> a06472a57304
Step 9/21 : FROM mcr.microsoft.com/powershell:lts-nanoserver-${NANOSERVER_VERSION}
lts-nanoserver-1809: Pulling from powershell
af0153d864f1: Pulling fs layer
eb9aadfb9ebf: Pulling fs layer
698ae59174d5: Pulling fs layer
ca1ef8c236fc: Pulling fs layer
e3789ff6b796: Pulling fs layer
a9a89a12f338: Pulling fs layer
ccb30ac1f365: Pulling fs layer
0b2e07545fa8: Pulling fs layer
72408742a0d5: Pulling fs layer
ca1ef8c236fc: Waiting
0b2e07545fa8: Waiting
72408742a0d5: Waiting
e3789ff6b796: Waiting
a9a89a12f338: Waiting
ccb30ac1f365: Waiting
eb9aadfb9ebf: Verifying Checksum
eb9aadfb9ebf: Download complete
698ae59174d5: Verifying Checksum
698ae59174d5: Download complete
ca1ef8c236fc: Verifying Checksum
ca1ef8c236fc: Download complete
e3789ff6b796: Verifying Checksum
e3789ff6b796: Download complete
af0153d864f1: Verifying Checksum
af0153d864f1: Download complete
ccb30ac1f365: Verifying Checksum
ccb30ac1f365: Download complete
0b2e07545fa8: Verifying Checksum
0b2e07545fa8: Download complete
72408742a0d5: Verifying Checksum
72408742a0d5: Download complete
a9a89a12f338: Verifying Checksum
a9a89a12f338: Download complete
af0153d864f1: Pull complete
eb9aadfb9ebf: Pull complete
698ae59174d5: Pull complete
ca1ef8c236fc: Pull complete
e3789ff6b796: Pull complete
a9a89a12f338: Pull complete
ccb30ac1f365: Pull complete
0b2e07545fa8: Pull complete
72408742a0d5: Pull complete
Digest: sha256:41643cdb0d064e3a282badd7c1568e8ffa904536a3e95116ce353696a5d1d471
Status: Downloaded newer image for mcr.microsoft.com/powershell:lts-nanoserver-1809
---> d57d56a6e450
Step 10/21 : SHELL ["pwsh", "-NoLogo", "-Command", "$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"]
---> Running in 494b3abb7438
Removing intermediate container 494b3abb7438
---> ad61f45fa627
Step 11/21 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Running in 3c3062b996af
Removing intermediate container 3c3062b996af
---> 4362e1e85c5d
Step 12/21 : LABEL description="A Docker image to deploy the Antrea CNI."
---> Running in 950d9303ee7c
Removing intermediate container 950d9303ee7c
---> a6d107b46ad6
Step 13/21 : USER ContainerAdministrator
---> Running in 08026f87f27e
Removing intermediate container 08026f87f27e
---> 5ed0dd7211b1
Step 14/21 : RUN mkdir -Force C:\k\antrea\bin
---> Running in 53264a32f2da
Directory: C:\k\antrea
[32;1mMode LastWriteTime Length Name[0m
[32;1m---- ------------- ------ ----[0m
d---- 3/8/2023 8:54 PM bin
Removing intermediate container 53264a32f2da
---> 744d452bdcef
Step 15/21 : COPY --from=antrea-build-windows /opt/cni/bin /k/antrea/cni
---> da2f7aea09d2
Step 16/21 : COPY --from=antrea-build-windows /antrea/build/images/scripts/Install-WindowsCNI.ps1 /k/antrea/
---> b3509d1238ca
Step 17/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-agent.exe /k/antrea/bin/
---> c75b95fbe770
Step 18/21 : COPY --from=antrea-build-windows /antrea/bin/antctl.exe /k/antrea/bin/antctl.exe
---> a6560ddef3ad
Step 19/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-cni.exe /k/antrea/cni/antrea.exe
---> ddf94081e150
Step 20/21 : RUN mkdir C:\k\antrea\utils
---> Running in 5e156644b55e
Directory: C:\k\antrea
[32;1mMode LastWriteTime Length Name[0m
[32;1m---- ------------- ------ ----[0m
d---- 3/8/2023 8:55 PM utils
Removing intermediate container 5e156644b55e
---> db43d08befea
Step 21/21 : COPY --from=antrea-build-windows /wins/wins.exe /k/antrea/utils/wins.exe
---> f84ef4d5fd9d
[Warning] One or more build-args [CNI_BINARIES_VERSION GO_VERSION] were not consumed
Successfully built f84ef4d5fd9d
Successfully tagged antrea/antrea-windows:v1.11.0-dev-75a53913.dirty
docker tag antrea/antrea-windows:v1.11.0-dev-75a53913.dirty antrea/antrea-windows
==== Reverting Windows VM a-ms-0008-win-1 =====
Windows VM a-ms-0008-win-1 powered on
The command completed successfully.
projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.29
projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0
projects.registry.vmware.com/antrea/e2eteam-pause:3.2
mcr.microsoft.com/windows/servercore/iis:latest
The image projects.registry.vmware.com/antrea/antrea-windows:latest already exists, renaming the old one with ID sha256:4653103584af17d23287e7ce178d037e588f8188506402323dc0dff943264b2b to empty string
Loaded image: projects.registry.vmware.com/antrea/antrea-windows:latest
====== Running Antrea e2e Tests ======
customresourcedefinition.apiextensions.k8s.io/antreaagentinfos.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/antreacontrollerinfos.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/clustergroups.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/clusternetworkpolicies.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/egresses.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalentities.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalippools.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalnodes.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/supportbundlecollections.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/tiers.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/traceflows.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/trafficcontrols.crd.antrea.io created
serviceaccount/antrea-agent created
serviceaccount/antctl created
serviceaccount/antrea-controller created
secret/antrea-agent-service-account-token created
secret/antctl-service-account-token created
configmap/antrea-config created
customresourcedefinition.apiextensions.k8s.io/groups.crd.antrea.io created
clusterrole.rbac.authorization.k8s.io/antrea-agent created
clusterrole.rbac.authorization.k8s.io/antctl created
clusterrole.rbac.authorization.k8s.io/antrea-cluster-identity-reader created
clusterrole.rbac.authorization.k8s.io/antrea-controller created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-view created
clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-view created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-view created
clusterrolebinding.rbac.authorization.k8s.io/antrea-agent created
clusterrolebinding.rbac.authorization.k8s.io/antctl created
clusterrolebinding.rbac.authorization.k8s.io/antrea-controller created
service/antrea created
daemonset.apps/antrea-agent created
deployment.apps/antrea-controller created
apiservice.apiregistration.k8s.io/v1beta2.controlplane.antrea.io created
apiservice.apiregistration.k8s.io/v1beta1.system.antrea.io created
apiservice.apiregistration.k8s.io/v1alpha1.stats.antrea.io created
mutatingwebhookconfiguration.admissionregistration.k8s.io/crdmutator.antrea.io created
validatingwebhookconfiguration.admissionregistration.k8s.io/crdvalidator.antrea.io created
configmap/antrea-agent-windows-kht6m7hthm created
configmap/antrea-windows-config-db2774h9dt created
daemonset.apps/antrea-agent-windows created
deployment.apps/coredns restarted
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 of 2 updated replicas are available...
deployment "coredns" successfully rolled out
deployment "antrea-controller" successfully rolled out
daemon set "antrea-agent" successfully rolled out
Waiting for daemon set "antrea-agent-windows" rollout to finish: 0 of 2 updated pods are available...
Waiting for daemon set "antrea-agent-windows" rollout to finish: 1 of 2 updated pods are available...
daemon set "antrea-agent-windows" successfully rolled out
Name InterfaceDescription ifIndex Status MacAddress LinkSpeed
---- -------------------- ------- ------ ---------- ---------
br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-BF-B1 10 Gbps
Name InterfaceDescription ifIndex Status MacAddress LinkSpeed
---- -------------------- ------- ------ ---------- ---------
br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-B7-EE 10 Gbps
=== Generate ssh-config ===
Generating ssh-config for Node a-ms-0008-0
Generating ssh-config for Node a-ms-0008-1
Generating ssh-config for Node a-ms-0008-win-0
Generating ssh-config for Node a-ms-0008-win-1
====== Run test with e2e test ======
...
``` | 2.0 | windows e2e test used stale image on one Node - **Describe the bug**
<!--
A clear and concise description of what the bug is.
If you believe this bug is a security issue, please don't use this template and follow our [security guidelines](/SECURITY.md)
-->
When I was debugging an issue on a windows testbed, I found one antrea-agent-windows Pod ran an old image. The correct version should be v1.11.0-dev-75a53913.dirty:
```
# antrea-agent-windows-nrpvd
I0308 21:10:38.550971 6844 agent.go:99] Starting Antrea agent (version v1.11.0-dev-75a53913.dirty)
# antrea-agent-windows-ztlsr
I0308 21:09:47.755164 1484 agent.go:92] Starting Antrea agent (version v1.7.0-dev-2a37aec1.dirty)
```
Testbed deployment log:
```
[antrea-windows-e2e-proxyall-for-pull-request] $ /bin/sh -xe /tmp/jenkins8184209255165314312.sh
+ set -e
+ head -n1 ci/docker-registry
+ DOCKER_REGISTRY=projects.registry.vmware.com
+ sed -i /skipIfNotIPv4Cluster(t)/a\ \ \ \ \ \ \ \ skipIfHasWindowsNodes(t) test/e2e/nodeportlocal_test.go
+ ./ci/jenkins/test.sh --testcase windows-e2e --registry projects.registry.vmware.com --proxyall
===== Clean up stale files & folders older than 7 days under /tmp =====
====== Cleanup Antrea Installation ======
No resources found in antrea-test namespace.
====== Building Antrea for the Following Commit ======
commit 75a5391315ac5d89f2a1e57f3cecbca3f816ade5
Merge: 32485497 18ddb089
Author: Kumar Atish <atish.iaf@gmail.com>
Date: Wed Mar 8 23:03:12 2023 +0530
Merge 18ddb08974afc6d051c489fb87890a495841cb56 into 324854979d383c11e83e49b5c7e88d1caec6ebdd
7 1 pkg/agent/apiserver/handlers/memberlist/handler.go
62 26 pkg/agent/apiserver/handlers/memberlist/handler_test.go
5 1 pkg/support/dump.go
Deleted Images:
untagged: antrea/cni-binaries:v1.1.1
untagged: projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
untagged: projects.registry.vmware.com/antrea/cni-binaries@sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219
deleted: sha256:bfaa431b732dd3d12776c98d0e684f65e9cf77c481b9e35496b3e1b02e25da05
deleted: sha256:fe0b72db62131760671074f192f8c182a0c22c6fa2292316123ebdf71ddb77f5
deleted: sha256:55139abd5e2a8db4f73283a7f50ed7b3b1d81d4d2956ec119a3d1f03c2eaeb4d
deleted: sha256:7e7fc777a83c92df7fb369f3b2a61189218f567944bbf9269f69453372c109d9
deleted: sha256:25bb6906eea3f717d8295b99a4d69538624cedafa62c04bf7dff0b4ed97214d3
deleted: sha256:f6d6f546f40138b7b433a17e74af250fa533cbd34e24783d0ceef0a1b7fc7b64
deleted: sha256:67ce281807736e142e8151f56dd4a2e387d9389449c170eb7405c04761479ed4
untagged: ubuntu:22.04
untagged: projects.registry.vmware.com/antrea/ubuntu:22.04
untagged: projects.registry.vmware.com/antrea/ubuntu@sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b
untagged: e2eteam/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1
deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4
deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9
deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3
deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88
deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac
deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf
deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463
deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68
deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc
deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac
untagged: antrea/openvswitch-debs:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch-debs@sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668
deleted: sha256:fd116dd5d80d4f166894a08e1832afaf52fbbd6367dad185199094c996f62301
deleted: sha256:81f789fce63bda6d07e049856d386c61a636528699c76d022708ddee445b4b2e
deleted: sha256:6b8158b5883be250d6f351957fb43c33a4e4c4b63898aa7a2a86522e0f0dc523
deleted: sha256:2eb4d98e07106613c35d3a505adcf490871ceeb3de9ce520c3b5ed1ab757ebf1
deleted: sha256:c960d4cb5b1e6437b012bf69ee311156c1b63ce80dae9bec4277b26b9947b742
deleted: sha256:e2b36c6bf533a216accfb28a94c4acb21ee631cab9b8eab1d723e65ce176d71a
deleted: sha256:a68b22d0c835019f3f34a2eeabf873611708513b616483c091d94673ed9cc329
untagged: nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101
deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478
deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f
deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a
deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282
deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81
untagged: antrea/openvswitch:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/openvswitch@sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83
untagged: antrea/base-ubuntu:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
untagged: projects.registry.vmware.com/antrea/base-ubuntu@sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d
Total reclaimed space: 773.4MB
BUILD_TAG: antrea-v1.11
22.04: Pulling from antrea/ubuntu
Digest: sha256:e99601e2804a9cc80cc3df6b7ccbc9f230d27a7d454e88408b4f5aeae38e138b
Status: Downloaded newer image for projects.registry.vmware.com/antrea/ubuntu:22.04
projects.registry.vmware.com/antrea/ubuntu:22.04
1.19: Pulling from antrea/golang
Digest: sha256:6687844637736cc26e4a10386abb7b482e5d231574be0bef9260a255da1f4775
Status: Image is up to date for projects.registry.vmware.com/antrea/golang:1.19
projects.registry.vmware.com/antrea/golang:1.19
antrea-v1.11: Pulling from antrea/openvswitch-debs
76769433fd8a: Already exists
57d34a411582: Pulling fs layer
7dad3cb1dcc7: Pulling fs layer
4f21db537580: Pulling fs layer
7dad3cb1dcc7: Verifying Checksum
7dad3cb1dcc7: Download complete
4f21db537580: Verifying Checksum
4f21db537580: Download complete
57d34a411582: Verifying Checksum
57d34a411582: Download complete
57d34a411582: Pull complete
7dad3cb1dcc7: Pull complete
4f21db537580: Pull complete
Digest: sha256:5cb2acba674a5f57e32cb8b7a06c6c94a17f72222b2980814467c9c4c7205668
Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
projects.registry.vmware.com/antrea/openvswitch-debs:antrea-v1.11
antrea-v1.11: Pulling from antrea/openvswitch
Digest: sha256:6cc5df6fa5af4f38521fcc0b0ea0a2694170c89151e74bb2b190a6000f913e83
Status: Downloaded newer image for projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
projects.registry.vmware.com/antrea/openvswitch:antrea-v1.11
v1.1.1: Pulling from antrea/cni-binaries
76769433fd8a: Already exists
c96ff6af3775: Pulling fs layer
247dd6786f33: Pulling fs layer
247dd6786f33: Verifying Checksum
247dd6786f33: Download complete
c96ff6af3775: Verifying Checksum
c96ff6af3775: Download complete
c96ff6af3775: Pull complete
247dd6786f33: Pull complete
Digest: sha256:71982abe6668257f869ff1918138838646e56be397bce07316310e614e9b5219
Status: Downloaded newer image for projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
projects.registry.vmware.com/antrea/cni-binaries:v1.1.1
antrea-v1.11: Pulling from antrea/base-ubuntu
Digest: sha256:bbe05b8f252dcb018871d73c4d1c84aa4124209198d6760f23b7a650236e201d
Status: Downloaded newer image for projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
projects.registry.vmware.com/antrea/base-ubuntu:antrea-v1.11
Sending build context to Docker daemon 24.06kB
Step 1/5 : FROM ubuntu:22.04 as ovs-debs
---> 74f2314a03de
Step 2/5 : ARG OVS_VERSION
---> Using cache
---> a68b22d0c835
Step 3/5 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev
---> Using cache
---> c960d4cb5b1e
Step 4/5 : COPY apply-patches.sh /
---> Using cache
---> 6b8158b5883b
Step 5/5 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch*
---> Using cache
---> fd116dd5d80d
Successfully built fd116dd5d80d
Successfully tagged antrea/openvswitch-debs:antrea-v1.11
Sending build context to Docker daemon 24.06kB
Step 1/11 : FROM ubuntu:22.04 as ovs-debs
---> 74f2314a03de
Step 2/11 : ARG OVS_VERSION
---> Using cache
---> a68b22d0c835
Step 3/11 : RUN apt-get update && DEBIAN_FRONTEND="noninteractive" apt-get install -y --no-install-recommends wget curl git ca-certificates build-essential fakeroot graphviz bzip2 autoconf automake debhelper dh-python dh-autoreconf libssl-dev libtool openssl procps python3-all python3-twisted python3-zope.interface python3-sphinx libunbound-dev
---> Using cache
---> c960d4cb5b1e
Step 4/11 : COPY apply-patches.sh /
---> Using cache
---> 6b8158b5883b
Step 5/11 : RUN wget -q -O - https://www.openvswitch.org/releases/openvswitch-$OVS_VERSION.tar.gz | tar xz -C /tmp && cd /tmp/openvswitch* && /apply-patches.sh && DEB_BUILD_OPTIONS='parallel=8 nocheck' fakeroot debian/rules binary && cd /tmp && mkdir ovs-debs && mv libopenvswitch_*.deb openvswitch-common_*.deb openvswitch-switch_*.deb python*-openvswitch_*.deb openvswitch-ipsec_*.deb ovs-debs/ && cd / && rm -rf /tmp/openvswitch*
---> Using cache
---> fd116dd5d80d
Step 6/11 : FROM ubuntu:22.04
---> 74f2314a03de
Step 7/11 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> 2873291cca14
Step 8/11 : LABEL description="A Docker image based on Ubuntu 22.04 which includes Open vSwitch built from source."
---> Using cache
---> 0e26ec9b7d9d
Step 9/11 : COPY --from=ovs-debs /tmp/ovs-debs/* /tmp/ovs-debs/
---> Using cache
---> 0c4db4416cce
Step 10/11 : COPY charon-logging.conf /tmp
---> Using cache
---> 79d23217a023
Step 11/11 : RUN apt-get update && apt-get install -y --no-install-recommends iptables logrotate libstrongswan-standard-plugins && (dpkg -i /tmp/ovs-debs/*.deb || apt-get -f -y --no-install-recommends install) && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && sed -i "/rotate /a\ #size 100M" /etc/logrotate.d/openvswitch-switch && sed -i "/^.*filelog.*{/r /tmp/charon-logging.conf" /etc/strongswan.d/charon-logging.conf && rm -rf /tmp/*
---> Using cache
---> 3852ac294105
Successfully built 3852ac294105
Successfully tagged antrea/openvswitch:antrea-v1.11
/var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request
Sending build context to Docker daemon 13.31kB
Step 1/7 : ARG BUILD_TAG
Step 2/7 : FROM ubuntu:22.04 as cni-binaries
---> 74f2314a03de
Step 3/7 : ARG CNI_BINARIES_VERSION
---> Using cache
---> 67ce28180773
Step 4/7 : ARG WHEREABOUTS_VERSION=v0.5.4
---> Using cache
---> f6d6f546f401
Step 5/7 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates
---> Using cache
---> 7e7fc777a83c
Step 6/7 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth"
---> Using cache
---> 55139abd5e2a
Step 7/7 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts
---> Using cache
---> bfaa431b732d
Successfully built bfaa431b732d
Successfully tagged antrea/cni-binaries:v1.1.1
Sending build context to Docker daemon 13.31kB
Step 1/15 : ARG BUILD_TAG
Step 2/15 : FROM ubuntu:22.04 as cni-binaries
---> 74f2314a03de
Step 3/15 : ARG CNI_BINARIES_VERSION
---> Using cache
---> 67ce28180773
Step 4/15 : ARG WHEREABOUTS_VERSION=v0.5.4
---> Using cache
---> f6d6f546f401
Step 5/15 : RUN apt-get update && apt-get install -y --no-install-recommends wget ca-certificates
---> Using cache
---> 7e7fc777a83c
Step 6/15 : ENV CNI_PLUGINS="./host-local ./loopback ./portmap ./bandwidth"
---> Using cache
---> 55139abd5e2a
Step 7/15 : RUN set -eux; dpkgArch="$(dpkg --print-architecture)"; case "${dpkgArch##*-}" in amd64) pluginsArch='amd64' ;; armhf) pluginsArch='arm' ;; arm64) pluginsArch='arm64' ;; *) pluginsArch=''; echo >&2; echo >&2 "unsupported architecture '$dpkgArch'"; echo >&2 ; exit 1 ;; esac; mkdir -p /opt/cni/bin; wget -q -O - [https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-${pluginsArch}-$CNI_BINARIES_VERSION.tgz](https://github.com/containernetworking/plugins/releases/download/$CNI_BINARIES_VERSION/cni-plugins-linux-$%7BpluginsArch%7D-$CNI_BINARIES_VERSION.tgz) | tar xz -C /opt/cni/bin $CNI_PLUGINS; wget -q -O /opt/cni/bin/whereabouts [https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-${pluginsArch}](https://github.com/k8snetworkplumbingwg/whereabouts/releases/download/$WHEREABOUTS_VERSION/whereabouts-$%7BpluginsArch%7D) && chmod +x /opt/cni/bin/whereabouts
---> Using cache
---> bfaa431b732d
Step 8/15 : FROM antrea/openvswitch:${BUILD_TAG}
---> 3852ac294105
Step 9/15 : ARG SURICATA_VERSION
---> Using cache
---> bafddc6af04e
Step 10/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> ec89fc76d92d
Step 11/15 : LABEL description="An Ubuntu based Docker base image for Antrea."
---> Using cache
---> dd8fb640925d
Step 12/15 : USER root
---> Using cache
---> 5c814d44d69b
Step 13/15 : ADD https://raw.githubusercontent.com/kubernetes-sigs/iptables-wrappers/9e6ce59c864623ea71a6f7d59c35fcb13a919b87/iptables-wrapper-installer.sh /iptables-wrapper-installer.sh
---> Using cache
---> 789847397959
Step 14/15 : RUN apt-get update && apt-get install -y --no-install-recommends ipset jq inotify-tools gpg-agent software-properties-common && add-apt-repository ppa:oisf/suricata-${SURICATA_VERSION} && apt-get update && apt-get install -y suricata && apt-get remove -y gpg-agent software-properties-common && apt-get autoremove -y && rm -rf /var/cache/apt/* /var/lib/apt/lists/* && chmod +x /iptables-wrapper-installer.sh && /iptables-wrapper-installer.sh
---> Using cache
---> 815257da934e
Step 15/15 : COPY --from=cni-binaries /opt/cni/bin /opt/cni/bin
---> Using cache
---> ad75b5563247
Successfully built ad75b5563247
Successfully tagged antrea/base-ubuntu:antrea-v1.11
/var/lib/jenkins/workspace/antrea-windows-e2e-proxyall-for-pull-request
===> Building Antrea bins and antrea/antrea-ubuntu Docker image <===
docker build -t antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.ubuntu --build-arg OVS_VERSION=2.17.3 --build-arg GO_VERSION=1.19 --build-arg BUILD_TAG=antrea-v1.11 .
Sending build context to Docker daemon 47.41MB
Step 1/15 : ARG GO_VERSION
Step 2/15 : ARG BUILD_TAG
Step 3/15 : FROM golang:${GO_VERSION} as antrea-build
---> ff3cd58379d9
Step 4/15 : WORKDIR /antrea
---> Running in 62533a8c8585
Removing intermediate container 62533a8c8585
---> 869906fadac2
Step 5/15 : COPY go.mod /antrea/go.mod
---> e087b2973e1c
Step 6/15 : RUN go mod download
---> Running in 6771daf43c3b
Removing intermediate container 6771daf43c3b
---> 41125c6f24ee
Step 7/15 : COPY . /antrea
---> de288ca18a4c
Step 8/15 : RUN make antrea-agent antrea-controller antrea-cni antctl-linux
---> Running in a3938237789d
GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-agent
GOOS=linux go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-controller
GOOS=linux CGO_ENABLED=0 go build -o /antrea/bin -ldflags '-X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni
Removing intermediate container a3938237789d
---> 848dec0d7aaf
Step 9/15 : RUN mv bin/antctl-linux bin/antctl
---> Running in b5313b457af3
Removing intermediate container b5313b457af3
---> f6bfc1318921
Step 10/15 : FROM antrea/base-ubuntu:${BUILD_TAG}
---> ad75b5563247
Step 11/15 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Using cache
---> 54aa6d3d57c4
Step 12/15 : LABEL description="The Docker image to deploy the Antrea CNI."
---> Using cache
---> 8161ff16a1cb
Step 13/15 : USER root
---> Using cache
---> ef8b9278df6d
Step 14/15 : COPY build/images/scripts/* /usr/local/bin/
---> Using cache
---> afd46dfce8c0
Step 15/15 : COPY --from=antrea-build /antrea/bin/* /usr/local/bin/
---> 2d3d7425fc7f
[Warning] One or more build-args [OVS_VERSION] were not consumed
Successfully built 2d3d7425fc7f
Successfully tagged antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty
docker tag antrea/antrea-ubuntu:v1.11.0-dev-75a53913.dirty antrea/antrea-ubuntu
====== Delivering Antrea to all Nodes ======
====== Updating yaml files to enable proxyAll ======
===== Pull necessary images on Control-Plane node =====
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/nginx:1.15-alpine
===== Deliver Antrea to Linux worker nodes and pull necessary images on worker nodes =====
sending incremental file list
antrea-ubuntu.tar
45,216 0% 0.00kB/s 0:00:00
158,854,896 31% 151.45MB/s 0:00:02
212,998,875 41% 101.60MB/s 0:00:02
282,112,496 55% 89.73MB/s 0:00:02
349,690,560 68% 83.42MB/s 0:00:01
402,000,296 78% 58.01MB/s 0:00:01
461,453,972 90% 59.28MB/s 0:00:00
511,165,952 100% 72.18MB/s 0:00:06 (xfr#1, to-chk=0/1)
sent 287,388,834 bytes received 158,320 bytes 33,829,076.94 bytes/sec
total size is 511,165,952 speedup is 1.78
Deleted Images:
untagged: k8s.gcr.io/e2e-test-images/agnhost:2.29
untagged: k8s.gcr.io/e2e-test-images/agnhost@sha256:1ae25df1a1985fa815f712b051f467dcd2d016d9efb74ca41ef385f9ed64ada9
deleted: sha256:3f4d3ab351ab4180a2c0fc655ba97263ba74fa353d41642b3e09bc16c231b735
deleted: sha256:9735e33cc1701e722d4ef42086090078232f4d45087ebea8a40de28b18eb190d
deleted: sha256:10f54c8640b3c21f6a7575ed67fa29c1811ed0a1404b0d398ab197ef2a55c46b
deleted: sha256:4104722c5b5696347fb190e2e341ac593a0dd3953ad63bf82c89e037f71da341
deleted: sha256:7abc3633883948ff206528cd56bada29a18222fd642bee4c76d67ab12e590113
deleted: sha256:41758cdb2ea620f0c2db17598ce779841c9d4dce6cb6c457b302e082b2d7771f
deleted: sha256:60e550c19b857d2261358edfa8b11cc43e53a076f3baea759553ce914889bd0d
deleted: sha256:aa75a9d85b48eba258a28d6975fb3db0393809cc290d3dab4d0bf2bdb97d1e79
deleted: sha256:891c281852b0ab60d0a307ea6c5fbd0697004ade3745e7adcfdd549391dd633f
deleted: sha256:33e8713114f88c8cb3f60c8a0a4aefe2500823b2fbbae05488b0185ba226caae
untagged: e2eteam/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost:2.13
untagged: projects.registry.vmware.com/antrea/agnhost@sha256:bbefb77f489d1aad44c3aa920cdaa528dcc83f94437007903e4624d42583eae1
deleted: sha256:445351f6b400f20e5f53ca483c4fecde679fd8bb418e3d5a6191ff9026c174a4
deleted: sha256:8c79532daf8d04a98acde1571304e68b2e606e6cbb603d24544952e5f1938da9
deleted: sha256:5240a558d0c6b17ad41a8a44db7e6ea97c7bd5a76f3bbf8f9023503a5bc0d5c3
deleted: sha256:981c4eb3e73ab104fb3698fe3be8a876ac6c5475ced6a9fb47dc47c6da2cbf88
deleted: sha256:c4f8e8a6c6d789fc9bfb6f11105337930de68b80e61321d479764b09e17906ac
deleted: sha256:a609cb4f69943879efca92b70fb11cf24b0befd95f9cbb18347bc1db18c483cf
deleted: sha256:e0555aa16ed6d0791cb627b11447c20c4efe48462e99b2c96155ad81216c6463
deleted: sha256:579ffe29f8f232edebd150d681b272551ccae8f9f8b5b28f902f6a59b4909c68
deleted: sha256:1b91ee0c583092e3516b64a708c3bdd7e27fb785adefbf8bea65b95305000dbc
deleted: sha256:721384ec99e56bc06202a738722bcb4b8254b9bbd71c43ab7ad0d9e773ced7ac
untagged: nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx:1.15-alpine
untagged: projects.registry.vmware.com/antrea/nginx@sha256:0fd68ec4b64b8dbb2bef1f1a5de9d47b658afd3635dc9c45bf0cbeac46e72101
deleted: sha256:dd025cdfe837e1c6395365870a491cf16bae668218edb07d85c626928a60e478
deleted: sha256:81fe0ee4a25b1cc9845049a29e70b3da5d465802030262fc378a6161594e9d0f
deleted: sha256:859b8cd9904ca7bb4adb5e56328df2a8c090c63dee93a089efba831fb4ada17a
deleted: sha256:9701f137fc094da480ccd272ca2aad1f788f4c0cbdc3c7ccbaa01d41411df282
deleted: sha256:f1b5933fe4b5f49bbe8258745cf396afe07e625bdab3168e364daf7c956b6b81
Total reclaimed space: 253.1MB
The image antrea/antrea-ubuntu:latest already exists, renaming the old one with ID sha256:568e9f48739ea51ab066586797f29459d1710bf9608f8ab7acf591d99b77bf73 to empty string
Loaded image: antrea/antrea-ubuntu:latest
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/nginx:1.15-alpine
k8s.gcr.io/e2e-test-images/agnhost:2.29
===== Deliver Antrea Windows to Windows worker nodes and pull necessary images on Windows worker nodes =====
==== Reverting Windows VM a-ms-0008-win-0 =====
Windows VM a-ms-0008-win-0 powered on
projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.29
projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0
projects.registry.vmware.com/antrea/e2eteam-pause:3.2
Error response from daemon: Get https://mcr.microsoft.com/v2/: x509: certificate has expired or is not yet valid
=== Build Windows on Windows Node===
1.19-nanoserver: Pulling from antrea/golang
5c9d6483dab1: Pulling fs layer
9982991b8208: Pulling fs layer
cb374f741579: Pulling fs layer
732dbfa4c42c: Pulling fs layer
64bbe06a1892: Pulling fs layer
15263a6e39d7: Pulling fs layer
d1ebeb495bd3: Pulling fs layer
460984e65c46: Pulling fs layer
109a63e20d99: Pulling fs layer
e3c52c6cbb5c: Pulling fs layer
64bbe06a1892: Waiting
15263a6e39d7: Waiting
d1ebeb495bd3: Waiting
460984e65c46: Waiting
109a63e20d99: Waiting
e3c52c6cbb5c: Waiting
732dbfa4c42c: Waiting
9982991b8208: Verifying Checksum
9982991b8208: Download complete
cb374f741579: Verifying Checksum
cb374f741579: Download complete
732dbfa4c42c: Verifying Checksum
732dbfa4c42c: Download complete
64bbe06a1892: Verifying Checksum
64bbe06a1892: Download complete
15263a6e39d7: Verifying Checksum
15263a6e39d7: Download complete
d1ebeb495bd3: Verifying Checksum
d1ebeb495bd3: Download complete
109a63e20d99: Download complete
e3c52c6cbb5c: Verifying Checksum
e3c52c6cbb5c: Download complete
5c9d6483dab1: Verifying Checksum
5c9d6483dab1: Download complete
460984e65c46: Verifying Checksum
460984e65c46: Download complete
5c9d6483dab1: Pull complete
9982991b8208: Pull complete
cb374f741579: Pull complete
732dbfa4c42c: Pull complete
64bbe06a1892: Pull complete
15263a6e39d7: Pull complete
d1ebeb495bd3: Pull complete
460984e65c46: Pull complete
109a63e20d99: Pull complete
e3c52c6cbb5c: Pull complete
Digest: sha256:d5843e74790da22bd3d84d098dcbe21c7318a7c7c4db6bf1de09cb193e64c513
Status: Downloaded newer image for projects.registry.vmware.com/antrea/golang:1.19-nanoserver
projects.registry.vmware.com/antrea/golang:1.19-nanoserver
===> Building Antrea bins and antrea/antrea-windows Docker image <===
docker build --pull -t antrea/antrea-windows:v1.11.0-dev-75a53913.dirty -f build/images/Dockerfile.build.windows --network host --build-arg GO_VERSION=1.19 --build-arg CNI_BINARIES_VERSION=v1.1.1 --build-arg NANOSERVER_VERSION=1809 --build-arg WIN_BUILD_TAG=caa8cac3d3 .
Sending build context to Docker daemon 558.6MB
Step 1/21 : ARG WIN_BUILD_TAG
Step 2/21 : ARG NANOSERVER_VERSION
Step 3/21 : FROM antrea/base-windows:${WIN_BUILD_TAG} as antrea-build-windows
caa8cac3d3: Pulling from antrea/base-windows
5ead999142ec: Pulling fs layer
3724c0f91ae7: Pulling fs layer
97a8c2800739: Pulling fs layer
c038a2fdb0b5: Pulling fs layer
5ef04caec17a: Pulling fs layer
8bca1755b69d: Pulling fs layer
3ebc2faa6c3e: Pulling fs layer
d85edb0b9152: Pulling fs layer
b343c39a6010: Pulling fs layer
b673c002e661: Pulling fs layer
99b72f4231b5: Pulling fs layer
d85edb0b9152: Waiting
3ebc2faa6c3e: Waiting
8bca1755b69d: Waiting
c038a2fdb0b5: Waiting
b343c39a6010: Waiting
5ef04caec17a: Waiting
b673c002e661: Waiting
99b72f4231b5: Waiting
97a8c2800739: Download complete
3724c0f91ae7: Download complete
5ead999142ec: Verifying Checksum
5ead999142ec: Download complete
c038a2fdb0b5: Download complete
5ef04caec17a: Verifying Checksum
5ef04caec17a: Download complete
8bca1755b69d: Verifying Checksum
8bca1755b69d: Download complete
b343c39a6010: Verifying Checksum
b343c39a6010: Download complete
3ebc2faa6c3e: Verifying Checksum
3ebc2faa6c3e: Download complete
99b72f4231b5: Verifying Checksum
99b72f4231b5: Download complete
b673c002e661: Verifying Checksum
b673c002e661: Download complete
d85edb0b9152: Verifying Checksum
d85edb0b9152: Download complete
5ead999142ec: Pull complete
3724c0f91ae7: Pull complete
97a8c2800739: Pull complete
c038a2fdb0b5: Pull complete
5ef04caec17a: Pull complete
8bca1755b69d: Pull complete
3ebc2faa6c3e: Pull complete
d85edb0b9152: Pull complete
b343c39a6010: Pull complete
b673c002e661: Pull complete
99b72f4231b5: Pull complete
Digest: sha256:c6edebf1a80553a5d474b6663657e69bc5b2cf54b116a1759e4d053c609a575d
Status: Downloaded newer image for antrea/base-windows:caa8cac3d3
---> 79d3453071f3
Step 4/21 : WORKDIR /antrea
---> Running in 8019bfcc22c9
Removing intermediate container 8019bfcc22c9
---> 6f713bbf1027
Step 5/21 : COPY go.mod /antrea/go.mod
---> 30bbece0963c
Step 6/21 : RUN go mod download
---> Running in 7a09a9c079f4
Removing intermediate container 7a09a9c079f4
---> af9b4babd43e
Step 7/21 : COPY . /antrea
---> 60d82c848851
Step 8/21 : RUN sh -c 'make windows-bin'
---> Running in aab89dd36ca9
GOOS=windows CGO_ENABLED=0 go build -o C:/antrea/bin -ldflags ' -X antrea.io/antrea/pkg/version.Version=v1.11.0-dev -X antrea.io/antrea/pkg/version.GitSHA=75a53913 -X antrea.io/antrea/pkg/version.GitTreeState=dirty -X antrea.io/antrea/pkg/version.ReleaseStatus=unreleased' antrea.io/antrea/cmd/antrea-cni antrea.io/antrea/cmd/antrea-agent antrea.io/antrea/cmd/antctl
Removing intermediate container aab89dd36ca9
---> a06472a57304
Step 9/21 : FROM mcr.microsoft.com/powershell:lts-nanoserver-${NANOSERVER_VERSION}
lts-nanoserver-1809: Pulling from powershell
af0153d864f1: Pulling fs layer
eb9aadfb9ebf: Pulling fs layer
698ae59174d5: Pulling fs layer
ca1ef8c236fc: Pulling fs layer
e3789ff6b796: Pulling fs layer
a9a89a12f338: Pulling fs layer
ccb30ac1f365: Pulling fs layer
0b2e07545fa8: Pulling fs layer
72408742a0d5: Pulling fs layer
ca1ef8c236fc: Waiting
0b2e07545fa8: Waiting
72408742a0d5: Waiting
e3789ff6b796: Waiting
a9a89a12f338: Waiting
ccb30ac1f365: Waiting
eb9aadfb9ebf: Verifying Checksum
eb9aadfb9ebf: Download complete
698ae59174d5: Verifying Checksum
698ae59174d5: Download complete
ca1ef8c236fc: Verifying Checksum
ca1ef8c236fc: Download complete
e3789ff6b796: Verifying Checksum
e3789ff6b796: Download complete
af0153d864f1: Verifying Checksum
af0153d864f1: Download complete
ccb30ac1f365: Verifying Checksum
ccb30ac1f365: Download complete
0b2e07545fa8: Verifying Checksum
0b2e07545fa8: Download complete
72408742a0d5: Verifying Checksum
72408742a0d5: Download complete
a9a89a12f338: Verifying Checksum
a9a89a12f338: Download complete
af0153d864f1: Pull complete
eb9aadfb9ebf: Pull complete
698ae59174d5: Pull complete
ca1ef8c236fc: Pull complete
e3789ff6b796: Pull complete
a9a89a12f338: Pull complete
ccb30ac1f365: Pull complete
0b2e07545fa8: Pull complete
72408742a0d5: Pull complete
Digest: sha256:41643cdb0d064e3a282badd7c1568e8ffa904536a3e95116ce353696a5d1d471
Status: Downloaded newer image for mcr.microsoft.com/powershell:lts-nanoserver-1809
---> d57d56a6e450
Step 10/21 : SHELL ["pwsh", "-NoLogo", "-Command", "$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"]
---> Running in 494b3abb7438
Removing intermediate container 494b3abb7438
---> ad61f45fa627
Step 11/21 : LABEL maintainer="Antrea <projectantrea-dev@googlegroups.com>"
---> Running in 3c3062b996af
Removing intermediate container 3c3062b996af
---> 4362e1e85c5d
Step 12/21 : LABEL description="A Docker image to deploy the Antrea CNI."
---> Running in 950d9303ee7c
Removing intermediate container 950d9303ee7c
---> a6d107b46ad6
Step 13/21 : USER ContainerAdministrator
---> Running in 08026f87f27e
Removing intermediate container 08026f87f27e
---> 5ed0dd7211b1
Step 14/21 : RUN mkdir -Force C:\k\antrea\bin
---> Running in 53264a32f2da
Directory: C:\k\antrea
[32;1mMode LastWriteTime Length Name[0m
[32;1m---- ------------- ------ ----[0m
d---- 3/8/2023 8:54 PM bin
Removing intermediate container 53264a32f2da
---> 744d452bdcef
Step 15/21 : COPY --from=antrea-build-windows /opt/cni/bin /k/antrea/cni
---> da2f7aea09d2
Step 16/21 : COPY --from=antrea-build-windows /antrea/build/images/scripts/Install-WindowsCNI.ps1 /k/antrea/
---> b3509d1238ca
Step 17/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-agent.exe /k/antrea/bin/
---> c75b95fbe770
Step 18/21 : COPY --from=antrea-build-windows /antrea/bin/antctl.exe /k/antrea/bin/antctl.exe
---> a6560ddef3ad
Step 19/21 : COPY --from=antrea-build-windows /antrea/bin/antrea-cni.exe /k/antrea/cni/antrea.exe
---> ddf94081e150
Step 20/21 : RUN mkdir C:\k\antrea\utils
---> Running in 5e156644b55e
Directory: C:\k\antrea
[32;1mMode LastWriteTime Length Name[0m
[32;1m---- ------------- ------ ----[0m
d---- 3/8/2023 8:55 PM utils
Removing intermediate container 5e156644b55e
---> db43d08befea
Step 21/21 : COPY --from=antrea-build-windows /wins/wins.exe /k/antrea/utils/wins.exe
---> f84ef4d5fd9d
[Warning] One or more build-args [CNI_BINARIES_VERSION GO_VERSION] were not consumed
Successfully built f84ef4d5fd9d
Successfully tagged antrea/antrea-windows:v1.11.0-dev-75a53913.dirty
docker tag antrea/antrea-windows:v1.11.0-dev-75a53913.dirty antrea/antrea-windows
==== Reverting Windows VM a-ms-0008-win-1 =====
Windows VM a-ms-0008-win-1 powered on
The command completed successfully.
projects.registry.vmware.com/antrea/sigwindowstools-kube-proxy:v1.18.0
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.13
projects.registry.vmware.com/antrea/agnhost:2.29
projects.registry.vmware.com/antrea/e2eteam-jessie-dnsutils:1.0
projects.registry.vmware.com/antrea/e2eteam-pause:3.2
mcr.microsoft.com/windows/servercore/iis:latest
The image projects.registry.vmware.com/antrea/antrea-windows:latest already exists, renaming the old one with ID sha256:4653103584af17d23287e7ce178d037e588f8188506402323dc0dff943264b2b to empty string
Loaded image: projects.registry.vmware.com/antrea/antrea-windows:latest
====== Running Antrea e2e Tests ======
customresourcedefinition.apiextensions.k8s.io/antreaagentinfos.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/antreacontrollerinfos.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/clustergroups.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/clusternetworkpolicies.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/egresses.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalentities.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalippools.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/externalnodes.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/supportbundlecollections.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/tiers.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/traceflows.crd.antrea.io created
customresourcedefinition.apiextensions.k8s.io/trafficcontrols.crd.antrea.io created
serviceaccount/antrea-agent created
serviceaccount/antctl created
serviceaccount/antrea-controller created
secret/antrea-agent-service-account-token created
secret/antctl-service-account-token created
configmap/antrea-config created
customresourcedefinition.apiextensions.k8s.io/groups.crd.antrea.io created
clusterrole.rbac.authorization.k8s.io/antrea-agent created
clusterrole.rbac.authorization.k8s.io/antctl created
clusterrole.rbac.authorization.k8s.io/antrea-cluster-identity-reader created
clusterrole.rbac.authorization.k8s.io/antrea-controller created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-policies-view created
clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-traceflows-view created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-edit created
clusterrole.rbac.authorization.k8s.io/aggregate-antrea-clustergroups-view created
clusterrolebinding.rbac.authorization.k8s.io/antrea-agent created
clusterrolebinding.rbac.authorization.k8s.io/antctl created
clusterrolebinding.rbac.authorization.k8s.io/antrea-controller created
service/antrea created
daemonset.apps/antrea-agent created
deployment.apps/antrea-controller created
apiservice.apiregistration.k8s.io/v1beta2.controlplane.antrea.io created
apiservice.apiregistration.k8s.io/v1beta1.system.antrea.io created
apiservice.apiregistration.k8s.io/v1alpha1.stats.antrea.io created
mutatingwebhookconfiguration.admissionregistration.k8s.io/crdmutator.antrea.io created
validatingwebhookconfiguration.admissionregistration.k8s.io/crdvalidator.antrea.io created
configmap/antrea-agent-windows-kht6m7hthm created
configmap/antrea-windows-config-db2774h9dt created
daemonset.apps/antrea-agent-windows created
deployment.apps/coredns restarted
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 out of 2 new replicas have been updated...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 old replicas are pending termination...
Waiting for deployment "coredns" rollout to finish: 1 of 2 updated replicas are available...
deployment "coredns" successfully rolled out
deployment "antrea-controller" successfully rolled out
daemon set "antrea-agent" successfully rolled out
Waiting for daemon set "antrea-agent-windows" rollout to finish: 0 of 2 updated pods are available...
Waiting for daemon set "antrea-agent-windows" rollout to finish: 1 of 2 updated pods are available...
daemon set "antrea-agent-windows" successfully rolled out
Name InterfaceDescription ifIndex Status MacAddress LinkSpeed
---- -------------------- ------- ------ ---------- ---------
br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-BF-B1 10 Gbps
Name InterfaceDescription ifIndex Status MacAddress LinkSpeed
---- -------------------- ------- ------ ---------- ---------
br-int Hyper-V Virtual Ethernet Adapter #4 19 Up 00-50-56-B1-B7-EE 10 Gbps
=== Generate ssh-config ===
Generating ssh-config for Node a-ms-0008-0
Generating ssh-config for Node a-ms-0008-1
Generating ssh-config for Node a-ms-0008-win-0
Generating ssh-config for Node a-ms-0008-win-1
====== Run test with e2e test ======
...
``` | test | windows test used stale image on one node describe the bug a clear and concise description of what the bug is if you believe this bug is a security issue please don t use this template and follow our security md when i was debugging an issue on a windows testbed i found one antrea agent windows pod ran an old image the correct version should be dev dirty antrea agent windows nrpvd agent go starting antrea agent version dev dirty antrea agent windows ztlsr agent go starting antrea agent version dev dirty testbed deployment log bin sh xe tmp sh set e head ci docker registry docker registry projects registry vmware com sed i t a skipifhaswindowsnodes t test nodeportlocal test go ci jenkins test sh testcase windows registry projects registry vmware com proxyall clean up stale files folders older than days under tmp cleanup antrea installation no resources found in antrea test namespace building antrea for the following commit commit merge author kumar atish date wed mar merge into pkg agent apiserver handlers memberlist handler go pkg agent apiserver handlers memberlist handler test go pkg support dump go deleted images untagged antrea cni binaries untagged projects registry vmware com antrea cni binaries untagged projects registry vmware com antrea cni binaries deleted deleted deleted deleted deleted deleted deleted untagged ubuntu untagged projects registry vmware com antrea ubuntu untagged projects registry vmware com antrea ubuntu untagged agnhost untagged projects registry vmware com antrea agnhost untagged projects registry vmware com antrea agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged antrea openvswitch debs antrea untagged projects registry vmware com antrea openvswitch debs antrea untagged projects registry vmware com antrea openvswitch debs deleted deleted deleted deleted deleted deleted deleted untagged nginx alpine untagged projects registry vmware com antrea nginx alpine untagged projects registry vmware com antrea nginx deleted deleted deleted deleted deleted untagged antrea openvswitch antrea untagged projects registry vmware com antrea openvswitch antrea untagged projects registry vmware com antrea openvswitch untagged antrea base ubuntu antrea untagged projects registry vmware com antrea base ubuntu antrea untagged projects registry vmware com antrea base ubuntu total reclaimed space build tag antrea pulling from antrea ubuntu digest status downloaded newer image for projects registry vmware com antrea ubuntu projects registry vmware com antrea ubuntu pulling from antrea golang digest status image is up to date for projects registry vmware com antrea golang projects registry vmware com antrea golang antrea pulling from antrea openvswitch debs already exists pulling fs layer pulling fs layer pulling fs layer verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea openvswitch debs antrea projects registry vmware com antrea openvswitch debs antrea antrea pulling from antrea openvswitch digest status downloaded newer image for projects registry vmware com antrea openvswitch antrea projects registry vmware com antrea openvswitch antrea pulling from antrea cni binaries already exists pulling fs layer pulling fs layer verifying checksum download complete verifying checksum download complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea cni binaries projects registry vmware com antrea cni binaries antrea pulling from antrea base ubuntu digest status downloaded newer image for projects registry vmware com antrea base ubuntu antrea projects registry vmware com antrea base ubuntu antrea sending build context to docker daemon step from ubuntu as ovs debs step arg ovs version using cache step run apt get update debian frontend noninteractive apt get install y no install recommends wget curl git ca certificates build essential fakeroot graphviz autoconf automake debhelper dh python dh autoreconf libssl dev libtool openssl procps all twisted zope interface sphinx libunbound dev using cache step copy apply patches sh using cache step run wget q o tar xz c tmp cd tmp openvswitch apply patches sh deb build options parallel nocheck fakeroot debian rules binary cd tmp mkdir ovs debs mv libopenvswitch deb openvswitch common deb openvswitch switch deb python openvswitch deb openvswitch ipsec deb ovs debs cd rm rf tmp openvswitch using cache successfully built successfully tagged antrea openvswitch debs antrea sending build context to docker daemon step from ubuntu as ovs debs step arg ovs version using cache step run apt get update debian frontend noninteractive apt get install y no install recommends wget curl git ca certificates build essential fakeroot graphviz autoconf automake debhelper dh python dh autoreconf libssl dev libtool openssl procps all twisted zope interface sphinx libunbound dev using cache step copy apply patches sh using cache step run wget q o tar xz c tmp cd tmp openvswitch apply patches sh deb build options parallel nocheck fakeroot debian rules binary cd tmp mkdir ovs debs mv libopenvswitch deb openvswitch common deb openvswitch switch deb python openvswitch deb openvswitch ipsec deb ovs debs cd rm rf tmp openvswitch using cache step from ubuntu step label maintainer antrea using cache step label description a docker image based on ubuntu which includes open vswitch built from source using cache step copy from ovs debs tmp ovs debs tmp ovs debs using cache step copy charon logging conf tmp using cache step run apt get update apt get install y no install recommends iptables logrotate libstrongswan standard plugins dpkg i tmp ovs debs deb apt get f y no install recommends install rm rf var cache apt var lib apt lists sed i rotate a size etc logrotate d openvswitch switch sed i filelog r tmp charon logging conf etc strongswan d charon logging conf rm rf tmp using cache successfully built successfully tagged antrea openvswitch antrea var lib jenkins workspace antrea windows proxyall for pull request sending build context to docker daemon step arg build tag step from ubuntu as cni binaries step arg cni binaries version using cache step arg whereabouts version using cache step run apt get update apt get install y no install recommends wget ca certificates using cache step env cni plugins host local loopback portmap bandwidth using cache step run set eux dpkgarch dpkg print architecture case dpkgarch in pluginsarch armhf pluginsarch arm pluginsarch pluginsarch echo echo unsupported architecture dpkgarch echo exit esac mkdir p opt cni bin wget q o tar xz c opt cni bin cni plugins wget q o opt cni bin whereabouts chmod x opt cni bin whereabouts using cache successfully built successfully tagged antrea cni binaries sending build context to docker daemon step arg build tag step from ubuntu as cni binaries step arg cni binaries version using cache step arg whereabouts version using cache step run apt get update apt get install y no install recommends wget ca certificates using cache step env cni plugins host local loopback portmap bandwidth using cache step run set eux dpkgarch dpkg print architecture case dpkgarch in pluginsarch armhf pluginsarch arm pluginsarch pluginsarch echo echo unsupported architecture dpkgarch echo exit esac mkdir p opt cni bin wget q o tar xz c opt cni bin cni plugins wget q o opt cni bin whereabouts chmod x opt cni bin whereabouts using cache step from antrea openvswitch build tag step arg suricata version using cache step label maintainer antrea using cache step label description an ubuntu based docker base image for antrea using cache step user root using cache step add iptables wrapper installer sh using cache step run apt get update apt get install y no install recommends ipset jq inotify tools gpg agent software properties common add apt repository ppa oisf suricata suricata version apt get update apt get install y suricata apt get remove y gpg agent software properties common apt get autoremove y rm rf var cache apt var lib apt lists chmod x iptables wrapper installer sh iptables wrapper installer sh using cache step copy from cni binaries opt cni bin opt cni bin using cache successfully built successfully tagged antrea base ubuntu antrea var lib jenkins workspace antrea windows proxyall for pull request building antrea bins and antrea antrea ubuntu docker image docker build t antrea antrea ubuntu dev dirty f build images dockerfile build ubuntu build arg ovs version build arg go version build arg build tag antrea sending build context to docker daemon step arg go version step arg build tag step from golang go version as antrea build step workdir antrea running in removing intermediate container step copy go mod antrea go mod step run go mod download running in removing intermediate container step copy antrea step run make antrea agent antrea controller antrea cni antctl linux running in goos linux go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea agent goos linux go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea controller goos linux cgo enabled go build o antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea cni removing intermediate container step run mv bin antctl linux bin antctl running in removing intermediate container step from antrea base ubuntu build tag step label maintainer antrea using cache step label description the docker image to deploy the antrea cni using cache step user root using cache step copy build images scripts usr local bin using cache step copy from antrea build antrea bin usr local bin one or more build args were not consumed successfully built successfully tagged antrea antrea ubuntu dev dirty docker tag antrea antrea ubuntu dev dirty antrea antrea ubuntu delivering antrea to all nodes updating yaml files to enable proxyall pull necessary images on control plane node projects registry vmware com antrea agnhost projects registry vmware com antrea nginx alpine deliver antrea to linux worker nodes and pull necessary images on worker nodes sending incremental file list antrea ubuntu tar s s s s s s s s xfr to chk sent bytes received bytes bytes sec total size is speedup is deleted images untagged gcr io test images agnhost untagged gcr io test images agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged agnhost untagged projects registry vmware com antrea agnhost untagged projects registry vmware com antrea agnhost deleted deleted deleted deleted deleted deleted deleted deleted deleted deleted untagged nginx alpine untagged projects registry vmware com antrea nginx alpine untagged projects registry vmware com antrea nginx deleted deleted deleted deleted deleted total reclaimed space the image antrea antrea ubuntu latest already exists renaming the old one with id to empty string loaded image antrea antrea ubuntu latest projects registry vmware com antrea agnhost projects registry vmware com antrea nginx alpine gcr io test images agnhost deliver antrea windows to windows worker nodes and pull necessary images on windows worker nodes reverting windows vm a ms win windows vm a ms win powered on projects registry vmware com antrea sigwindowstools kube proxy projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea jessie dnsutils projects registry vmware com antrea pause error response from daemon get certificate has expired or is not yet valid build windows on windows node nanoserver pulling from antrea golang pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting waiting verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for projects registry vmware com antrea golang nanoserver projects registry vmware com antrea golang nanoserver building antrea bins and antrea antrea windows docker image docker build pull t antrea antrea windows dev dirty f build images dockerfile build windows network host build arg go version build arg cni binaries version build arg nanoserver version build arg win build tag sending build context to docker daemon step arg win build tag step arg nanoserver version step from antrea base windows win build tag as antrea build windows pulling from antrea base windows pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting waiting waiting download complete download complete verifying checksum download complete download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for antrea base windows step workdir antrea running in removing intermediate container step copy go mod antrea go mod step run go mod download running in removing intermediate container step copy antrea step run sh c make windows bin running in goos windows cgo enabled go build o c antrea bin ldflags x antrea io antrea pkg version version dev x antrea io antrea pkg version gitsha x antrea io antrea pkg version gittreestate dirty x antrea io antrea pkg version releasestatus unreleased antrea io antrea cmd antrea cni antrea io antrea cmd antrea agent antrea io antrea cmd antctl removing intermediate container step from mcr microsoft com powershell lts nanoserver nanoserver version lts nanoserver pulling from powershell pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer pulling fs layer waiting waiting waiting waiting waiting waiting verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete verifying checksum download complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for mcr microsoft com powershell lts nanoserver step shell running in removing intermediate container step label maintainer antrea running in removing intermediate container step label description a docker image to deploy the antrea cni running in removing intermediate container step user containeradministrator running in removing intermediate container step run mkdir force c k antrea bin running in directory c k antrea lastwritetime length name d pm bin removing intermediate container step copy from antrea build windows opt cni bin k antrea cni step copy from antrea build windows antrea build images scripts install windowscni k antrea step copy from antrea build windows antrea bin antrea agent exe k antrea bin step copy from antrea build windows antrea bin antctl exe k antrea bin antctl exe step copy from antrea build windows antrea bin antrea cni exe k antrea cni antrea exe step run mkdir c k antrea utils running in directory c k antrea lastwritetime length name d pm utils removing intermediate container step copy from antrea build windows wins wins exe k antrea utils wins exe one or more build args were not consumed successfully built successfully tagged antrea antrea windows dev dirty docker tag antrea antrea windows dev dirty antrea antrea windows reverting windows vm a ms win windows vm a ms win powered on the command completed successfully projects registry vmware com antrea sigwindowstools kube proxy projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea agnhost projects registry vmware com antrea jessie dnsutils projects registry vmware com antrea pause mcr microsoft com windows servercore iis latest the image projects registry vmware com antrea antrea windows latest already exists renaming the old one with id to empty string loaded image projects registry vmware com antrea antrea windows latest running antrea tests customresourcedefinition apiextensions io antreaagentinfos crd antrea io created customresourcedefinition apiextensions io antreacontrollerinfos crd antrea io created customresourcedefinition apiextensions io clustergroups crd antrea io created customresourcedefinition apiextensions io clusternetworkpolicies crd antrea io created customresourcedefinition apiextensions io egresses crd antrea io created customresourcedefinition apiextensions io externalentities crd antrea io created customresourcedefinition apiextensions io externalippools crd antrea io created customresourcedefinition apiextensions io externalnodes crd antrea io created customresourcedefinition apiextensions io ippools crd antrea io created customresourcedefinition apiextensions io networkpolicies crd antrea io created customresourcedefinition apiextensions io supportbundlecollections crd antrea io created customresourcedefinition apiextensions io tiers crd antrea io created customresourcedefinition apiextensions io traceflows crd antrea io created customresourcedefinition apiextensions io trafficcontrols crd antrea io created serviceaccount antrea agent created serviceaccount antctl created serviceaccount antrea controller created secret antrea agent service account token created secret antctl service account token created configmap antrea config created customresourcedefinition apiextensions io groups crd antrea io created clusterrole rbac authorization io antrea agent created clusterrole rbac authorization io antctl created clusterrole rbac authorization io antrea cluster identity reader created clusterrole rbac authorization io antrea controller created clusterrole rbac authorization io aggregate antrea policies edit created clusterrole rbac authorization io aggregate antrea policies view created clusterrole rbac authorization io aggregate traceflows edit created clusterrole rbac authorization io aggregate traceflows view created clusterrole rbac authorization io aggregate antrea clustergroups edit created clusterrole rbac authorization io aggregate antrea clustergroups view created clusterrolebinding rbac authorization io antrea agent created clusterrolebinding rbac authorization io antctl created clusterrolebinding rbac authorization io antrea controller created service antrea created daemonset apps antrea agent created deployment apps antrea controller created apiservice apiregistration io controlplane antrea io created apiservice apiregistration io system antrea io created apiservice apiregistration io stats antrea io created mutatingwebhookconfiguration admissionregistration io crdmutator antrea io created validatingwebhookconfiguration admissionregistration io crdvalidator antrea io created configmap antrea agent windows created configmap antrea windows config created daemonset apps antrea agent windows created deployment apps coredns restarted waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish out of new replicas have been updated waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish old replicas are pending termination waiting for deployment coredns rollout to finish of updated replicas are available deployment coredns successfully rolled out deployment antrea controller successfully rolled out daemon set antrea agent successfully rolled out waiting for daemon set antrea agent windows rollout to finish of updated pods are available waiting for daemon set antrea agent windows rollout to finish of updated pods are available daemon set antrea agent windows successfully rolled out name interfacedescription ifindex status macaddress linkspeed br int hyper v virtual ethernet adapter up bf gbps name interfacedescription ifindex status macaddress linkspeed br int hyper v virtual ethernet adapter up ee gbps generate ssh config generating ssh config for node a ms generating ssh config for node a ms generating ssh config for node a ms win generating ssh config for node a ms win run test with test | 1 |
80,598 | 7,751,829,219 | IssuesEvent | 2018-05-30 18:19:28 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | Test failure: GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd | test-run-uwp-coreclr | Opened on behalf of @Sunny-pu
The test `GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd` has failed.
Test Infrastructure Failure: Could not load file or assembly 'System.Memory, Version=4.1.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified.
Expected: True
Actual: False
Stack Trace:
Build : Master - 20180528.01 (Ready-To-Run Tests)
Failing configurations:
- ubuntu.1604.amd64
- x64
Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fr2r~2Fcli~2F/build/20180528.01/workItem/GC.Coverage.XUnitWrapper/analysis/xunit/GC_Coverage._delete_next_card_table_delete_next_card_table_~2F_delete_next_card_table_delete_next_card_table_cmd | 1.0 | Test failure: GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd - Opened on behalf of @Sunny-pu
The test `GC_Coverage._delete_next_card_table_delete_next_card_table_/_delete_next_card_table_delete_next_card_table_cmd` has failed.
Test Infrastructure Failure: Could not load file or assembly 'System.Memory, Version=4.1.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified.
Expected: True
Actual: False
Stack Trace:
Build : Master - 20180528.01 (Ready-To-Run Tests)
Failing configurations:
- ubuntu.1604.amd64
- x64
Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fr2r~2Fcli~2F/build/20180528.01/workItem/GC.Coverage.XUnitWrapper/analysis/xunit/GC_Coverage._delete_next_card_table_delete_next_card_table_~2F_delete_next_card_table_delete_next_card_table_cmd | test | test failure gc coverage delete next card table delete next card table delete next card table delete next card table cmd opened on behalf of sunny pu the test gc coverage delete next card table delete next card table delete next card table delete next card table cmd has failed test infrastructure failure could not load file or assembly system memory version culture neutral publickeytoken the system cannot find the file specified expected true actual false stack trace build master ready to run tests failing configurations ubuntu details | 1 |
132,796 | 10,764,165,103 | IssuesEvent | 2019-11-01 07:18:35 | a2000-erp-team/WEBERP | https://api.github.com/repos/a2000-erp-team/WEBERP | opened | Procurement-Admin-Item Account Group-Edit. Mapping for "Stock A/c" shows Stock Adjust A/c. Kindly correct because misleading, cannot find where to change the Stock A/c. mapping. | WEB ERP Testing By Katrina |
 | 1.0 | Procurement-Admin-Item Account Group-Edit. Mapping for "Stock A/c" shows Stock Adjust A/c. Kindly correct because misleading, cannot find where to change the Stock A/c. mapping. -
 | test | procurement admin item account group edit mapping for stock a c shows stock adjust a c kindly correct because misleading cannot find where to change the stock a c mapping | 1 |
302,862 | 26,169,264,911 | IssuesEvent | 2023-01-01 17:58:02 | LAION-AI/Open-Assistant | https://api.github.com/repos/LAION-AI/Open-Assistant | closed | Contract tests between API and Discord Bot | discord-bot testing | As per discussion in #180, the communication between the discord bot and API should be tested through a Contract test. This test should test that each call satisfies the OpenAPI schema inputs and endpoints. That way, changes to the endpoint contract will automatically be caught by these tests. | 1.0 | Contract tests between API and Discord Bot - As per discussion in #180, the communication between the discord bot and API should be tested through a Contract test. This test should test that each call satisfies the OpenAPI schema inputs and endpoints. That way, changes to the endpoint contract will automatically be caught by these tests. | test | contract tests between api and discord bot as per discussion in the communication between the discord bot and api should be tested through a contract test this test should test that each call satisfies the openapi schema inputs and endpoints that way changes to the endpoint contract will automatically be caught by these tests | 1 |
63,849 | 6,886,277,634 | IssuesEvent | 2017-11-21 18:53:56 | RetroWoW/PTR | https://api.github.com/repos/RetroWoW/PTR | closed | T2 Paladin 8/8 Bonus not scaling with spelldmg any more | Class - Paladin Requires Testing / Confirmation | **Description**:
With the new core the 8/8 set bonus does exactly the dmg that it says on the tooltip and doesn't scale with spelldmg any more
**Expected behavior**: Should scale
| 1.0 | T2 Paladin 8/8 Bonus not scaling with spelldmg any more - **Description**:
With the new core the 8/8 set bonus does exactly the dmg that it says on the tooltip and doesn't scale with spelldmg any more
**Expected behavior**: Should scale
| test | paladin bonus not scaling with spelldmg any more description with the new core the set bonus does exactly the dmg that it says on the tooltip and doesn t scale with spelldmg any more expected behavior should scale | 1 |
317,155 | 27,217,199,084 | IssuesEvent | 2023-02-20 23:35:57 | acikkaynak/deprem-yardim-frontend | https://api.github.com/repos/acikkaynak/deprem-yardim-frontend | closed | bug: Can't click scan field button when map layer window is open | bug discussion approved ios android p0 test-failed | ## Bug Definition
Can't click scan field button when map layer window is open
-- Please add discord contact information
** discord username: @afetharita#0001 **
## Bug environment
Describe the environment produces the bug. For example:
rc.afetharita.com
## Describe how you are producing the bug step by step
1. Go to '..RC.'rc.afetharita.com
2. Click map layer button
3. Click to scan field button
4. Bug appears
## Expected Behaviour
map layers should be closed and scan layer shold be clickable
## Screen shots
If possible, add screenshots to describe your bug.
## Desktop Information
- Operating System: Windows 10
- Browser chrome
- Version 110.0.5481.97 (Resmi Derleme) (64 bit) Sürümü
## Mobile Phone Information
- Devıce: [for example iPhone6]
- Operating System (with the version): [for example iOS8.1]
- Version [for example default browser, safari]
- Browser Version [for example 22]
## Additional Context
Add any other context about the bug here
| 1.0 | bug: Can't click scan field button when map layer window is open - ## Bug Definition
Can't click scan field button when map layer window is open
-- Please add discord contact information
** discord username: @afetharita#0001 **
## Bug environment
Describe the environment produces the bug. For example:
rc.afetharita.com
## Describe how you are producing the bug step by step
1. Go to '..RC.'rc.afetharita.com
2. Click map layer button
3. Click to scan field button
4. Bug appears
## Expected Behaviour
map layers should be closed and scan layer shold be clickable
## Screen shots
If possible, add screenshots to describe your bug.
## Desktop Information
- Operating System: Windows 10
- Browser chrome
- Version 110.0.5481.97 (Resmi Derleme) (64 bit) Sürümü
## Mobile Phone Information
- Devıce: [for example iPhone6]
- Operating System (with the version): [for example iOS8.1]
- Version [for example default browser, safari]
- Browser Version [for example 22]
## Additional Context
Add any other context about the bug here
| test | bug can t click scan field button when map layer window is open bug definition can t click scan field button when map layer window is open please add discord contact information discord username afetharita bug environment describe the environment produces the bug for example rc afetharita com describe how you are producing the bug step by step go to rc rc afetharita com click map layer button click to scan field button bug appears expected behaviour map layers should be closed and scan layer shold be clickable screen shots if possible add screenshots to describe your bug desktop information operating system windows browser chrome version resmi derleme bit sürümü mobile phone information devıce operating system with the version version browser version additional context add any other context about the bug here | 1 |
459,863 | 13,200,379,420 | IssuesEvent | 2020-08-14 08:07:10 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | play.google.com - site is not usable | browser-android-components engine-gecko priority-critical | <!-- @browser: Firefox Mobile 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/56599 -->
<!-- @extra_labels: browser-android-components -->
**URL**: https://play.google.com/store/apps/details?id%3Dcom.google.android.apps.maps%26referrer%3Dutm_source%253Dhelp-center%2526utm_medium%253Dhc-answer%2526utm_campaign%253Dhelp-center-mg-1250066
**Browser / Version**: Firefox Mobile 81.0
**Operating System**: Android
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
I am hacked
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200808213325</li><li>channel: nightly</li><li>hasTouchScreen: true</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | play.google.com - site is not usable - <!-- @browser: Firefox Mobile 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/56599 -->
<!-- @extra_labels: browser-android-components -->
**URL**: https://play.google.com/store/apps/details?id%3Dcom.google.android.apps.maps%26referrer%3Dutm_source%253Dhelp-center%2526utm_medium%253Dhc-answer%2526utm_campaign%253Dhelp-center-mg-1250066
**Browser / Version**: Firefox Mobile 81.0
**Operating System**: Android
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
I am hacked
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200808213325</li><li>channel: nightly</li><li>hasTouchScreen: true</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | play google com site is not usable url browser version firefox mobile operating system android tested another browser yes safari problem type site is not usable description unable to login steps to reproduce i am hacked browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️ | 0 |
113,705 | 17,150,880,143 | IssuesEvent | 2021-07-13 20:25:42 | snowdensb/braindump | https://api.github.com/repos/snowdensb/braindump | opened | CVE-2019-18797 (Medium) detected in node-sass-3.12.1.tgz, libsass3.3.6 | security vulnerability | ## CVE-2019-18797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-3.12.1.tgz</b>, <b>libsass3.3.6</b></p></summary>
<p>
<details><summary><b>node-sass-3.12.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz</a></p>
<p>Path to dependency file: braindump/package.json</p>
<p>Path to vulnerable library: braindump/node_modules/node-sass</p>
<p>
Dependency Hierarchy:
- gulp-sass-2.3.2.tgz (Root Library)
- :x: **node-sass-3.12.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/braindump/commit/815ae0afebcf867f02143f3ab9cf88b1d4dacdec">815ae0afebcf867f02143f3ab9cf88b1d4dacdec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.
<p>Publish Date: 2019-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p>
<p>Release Date: 2019-11-06</p>
<p>Fix Resolution: LibSass - 3.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"3.12.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:2.3.2;node-sass:3.12.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"LibSass - 3.6.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-18797","vulnerabilityDetails":"LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-18797 (Medium) detected in node-sass-3.12.1.tgz, libsass3.3.6 - ## CVE-2019-18797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-3.12.1.tgz</b>, <b>libsass3.3.6</b></p></summary>
<p>
<details><summary><b>node-sass-3.12.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-3.12.1.tgz</a></p>
<p>Path to dependency file: braindump/package.json</p>
<p>Path to vulnerable library: braindump/node_modules/node-sass</p>
<p>
Dependency Hierarchy:
- gulp-sass-2.3.2.tgz (Root Library)
- :x: **node-sass-3.12.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/braindump/commit/815ae0afebcf867f02143f3ab9cf88b1d4dacdec">815ae0afebcf867f02143f3ab9cf88b1d4dacdec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.
<p>Publish Date: 2019-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p>
<p>Release Date: 2019-11-06</p>
<p>Fix Resolution: LibSass - 3.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"3.12.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:2.3.2;node-sass:3.12.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"LibSass - 3.6.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-18797","vulnerabilityDetails":"LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_test | cve medium detected in node sass tgz cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file braindump package json path to vulnerable library braindump node modules node sass dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp sass node sass isminimumfixversionavailable true minimumfixversion libsass basebranches vulnerabilityidentifier cve vulnerabilitydetails libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp vulnerabilityurl | 0 |
772,401 | 27,120,677,639 | IssuesEvent | 2023-02-15 22:33:56 | Lightning-AI/lightning | https://api.github.com/repos/Lightning-AI/lightning | closed | QAT callback doesn't work with multi-input models or forward kwargs | bug help wanted 3rd party priority: 1 | ## 🐛 Bug
The `QuantizationAwareTraining` can only handle models whose forward signature looks like `forward(data)`. Adding additional positional or keyword arguments breaks the forward propagation.
## Please reproduce using the BoringModel
The `BoringModel` doesn't have the right forward signature to produce errors, but I included an equivalently simple model below
### To Reproduce
```python
import functools
import torch
import torch.nn as nn
class Simple(nn.Module):
def __init__(self):
super(Simple, self).__init__()
self.conv = nn.Conv2d(3, 16, kernel_size=3)
self.softmax = nn.Softmax(dim=3)
def forward(self, a, b, softmax=False):
"""Use 2 positional arguments and 1 keyword argument"""
feat_b = self.conv(b)
feat_a = self.conv(a)
combined = feat_a + feat_b
if softmax:
return self.softmax(combined)
else:
return combined
@property
def example_input_array(self):
return torch.randn(1, 3, 28, 28), torch.randn(1, 3, 28, 28)
def wrap_quantize_forward_context(model, func):
"""From QuantizationAwareTraining callback"""
@functools.wraps(func)
def wrapper(data):
data = model.quant(data)
data = func(data)
data = model.dequant(data)
return data
return wrapper
# Works fine without quantization
model = Simple()
input_sample = model.example_input_array
_ = model(*input_sample)
_ = model(*input_sample, softmax=True)
# Simulate the `on_fit_start` hook behavior
model.quant = torch.quantization.QuantStub()
model.dequant = torch.quantization.DeQuantStub()
model.forward = wrap_quantize_forward_context(model, model.forward)
_ = model(*input_sample)
_ = model(*input_sample, softmax=True)
```
The last two lines will raise the following exceptions
```
TypeError: wrapper() takes 1 positional argument but 2 were given
TypeError: wrapper() got an unexpected keyword argument 'softmax'
```
### Expected behavior
Forward prop should work fine for both the quantized and unquantized model
### Environment
* CUDA:
- GPU: GeForce GTX 1060 with Max-Q Design
- available: True
- version: 10.2
* Packages:
- numpy: 1.19.2
- pyTorch_debug: False
- pyTorch_version: 1.8.1+cu102
- pytorch-lightning: 1.3.0
- tqdm: 4.59.0
* System:
- OS: Linux
- architecture: 64bit
- processor: x86_64
- python: 3.6.13
- version: #148-Ubuntu SMP Sat May 8 02:33:43 UTC 2021
### Additional context
The solution to the minimal example is to make the decoration wrapper more generic
```python
def wrap_quantize_forward_context(model, func):
"""From QuantizationAwareTraining callback"""
@functools.wraps(func)
def wrapper(*data, **kwargs):
data = model.quant(data)
data = func(*data, **kwargs)
data = model.dequant(data)
return data
return wrapper
```
However, I notice that `callbacks/quantization.py` also includes a second decorator `wrap_qat_forward_context`, and I am not exactly clear on the intent behind having two separate decorators. I see that `wrap_quantize_forward_context` is only used in the `on_fit_end` hook. Any insight here @Borda?
I can make a PR for the above solution, but first I tested it in an actual training by patching my local install
```python
def wrap_qat_forward_context(
quant_cb,
model: 'pl.LightningModule',
func: Callable,
trigger_condition: Optional[Union[Callable, int]] = None
) -> Callable:
@functools.wraps(func)
def wrapper(*data, **kwargs) -> Any: # <-- Changes to decoration signature
_is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer)
_is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition
_quant_run = trigger_condition is None or _is_func_true or _is_count_true
if _quant_run:
quant_cb._forward_calls += 1
data = model.quant(data) # <-- Triggers AttributeError 'tuple' object has no attribute 'detach'
data = func(*data, **kwargs)
if _quant_run:
data = model.dequant(data)
return data
return wrapper
```
So I tried also expanding the `data` before passing to `model.quant` but this goes back to a `TypeError`
```python
def wrap_qat_forward_context(
quant_cb,
model: 'pl.LightningModule',
func: Callable,
trigger_condition: Optional[Union[Callable, int]] = None
) -> Callable:
@functools.wraps(func)
def wrapper(*data, **kwargs) -> Any:
_is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer)
_is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition
_quant_run = trigger_condition is None or _is_func_true or _is_count_true
if _quant_run:
quant_cb._forward_calls += 1
data = model.quant(*data) # <-- Attempted tuple unpacking, get TypeError: forward() takes 2 positional arguments but 3 were given
data = func(*data, **kwargs)
if _quant_run:
data = model.dequant(*data)
return data
return wrapper
``` | 1.0 | QAT callback doesn't work with multi-input models or forward kwargs - ## 🐛 Bug
The `QuantizationAwareTraining` can only handle models whose forward signature looks like `forward(data)`. Adding additional positional or keyword arguments breaks the forward propagation.
## Please reproduce using the BoringModel
The `BoringModel` doesn't have the right forward signature to produce errors, but I included an equivalently simple model below
### To Reproduce
```python
import functools
import torch
import torch.nn as nn
class Simple(nn.Module):
def __init__(self):
super(Simple, self).__init__()
self.conv = nn.Conv2d(3, 16, kernel_size=3)
self.softmax = nn.Softmax(dim=3)
def forward(self, a, b, softmax=False):
"""Use 2 positional arguments and 1 keyword argument"""
feat_b = self.conv(b)
feat_a = self.conv(a)
combined = feat_a + feat_b
if softmax:
return self.softmax(combined)
else:
return combined
@property
def example_input_array(self):
return torch.randn(1, 3, 28, 28), torch.randn(1, 3, 28, 28)
def wrap_quantize_forward_context(model, func):
"""From QuantizationAwareTraining callback"""
@functools.wraps(func)
def wrapper(data):
data = model.quant(data)
data = func(data)
data = model.dequant(data)
return data
return wrapper
# Works fine without quantization
model = Simple()
input_sample = model.example_input_array
_ = model(*input_sample)
_ = model(*input_sample, softmax=True)
# Simulate the `on_fit_start` hook behavior
model.quant = torch.quantization.QuantStub()
model.dequant = torch.quantization.DeQuantStub()
model.forward = wrap_quantize_forward_context(model, model.forward)
_ = model(*input_sample)
_ = model(*input_sample, softmax=True)
```
The last two lines will raise the following exceptions
```
TypeError: wrapper() takes 1 positional argument but 2 were given
TypeError: wrapper() got an unexpected keyword argument 'softmax'
```
### Expected behavior
Forward prop should work fine for both the quantized and unquantized model
### Environment
* CUDA:
- GPU: GeForce GTX 1060 with Max-Q Design
- available: True
- version: 10.2
* Packages:
- numpy: 1.19.2
- pyTorch_debug: False
- pyTorch_version: 1.8.1+cu102
- pytorch-lightning: 1.3.0
- tqdm: 4.59.0
* System:
- OS: Linux
- architecture: 64bit
- processor: x86_64
- python: 3.6.13
- version: #148-Ubuntu SMP Sat May 8 02:33:43 UTC 2021
### Additional context
The solution to the minimal example is to make the decoration wrapper more generic
```python
def wrap_quantize_forward_context(model, func):
"""From QuantizationAwareTraining callback"""
@functools.wraps(func)
def wrapper(*data, **kwargs):
data = model.quant(data)
data = func(*data, **kwargs)
data = model.dequant(data)
return data
return wrapper
```
However, I notice that `callbacks/quantization.py` also includes a second decorator `wrap_qat_forward_context`, and I am not exactly clear on the intent behind having two separate decorators. I see that `wrap_quantize_forward_context` is only used in the `on_fit_end` hook. Any insight here @Borda?
I can make a PR for the above solution, but first I tested it in an actual training by patching my local install
```python
def wrap_qat_forward_context(
quant_cb,
model: 'pl.LightningModule',
func: Callable,
trigger_condition: Optional[Union[Callable, int]] = None
) -> Callable:
@functools.wraps(func)
def wrapper(*data, **kwargs) -> Any: # <-- Changes to decoration signature
_is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer)
_is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition
_quant_run = trigger_condition is None or _is_func_true or _is_count_true
if _quant_run:
quant_cb._forward_calls += 1
data = model.quant(data) # <-- Triggers AttributeError 'tuple' object has no attribute 'detach'
data = func(*data, **kwargs)
if _quant_run:
data = model.dequant(data)
return data
return wrapper
```
So I tried also expanding the `data` before passing to `model.quant` but this goes back to a `TypeError`
```python
def wrap_qat_forward_context(
quant_cb,
model: 'pl.LightningModule',
func: Callable,
trigger_condition: Optional[Union[Callable, int]] = None
) -> Callable:
@functools.wraps(func)
def wrapper(*data, **kwargs) -> Any:
_is_func_true = isinstance(trigger_condition, Callable) and trigger_condition(model.trainer)
_is_count_true = isinstance(trigger_condition, int) and quant_cb._forward_calls < trigger_condition
_quant_run = trigger_condition is None or _is_func_true or _is_count_true
if _quant_run:
quant_cb._forward_calls += 1
data = model.quant(*data) # <-- Attempted tuple unpacking, get TypeError: forward() takes 2 positional arguments but 3 were given
data = func(*data, **kwargs)
if _quant_run:
data = model.dequant(*data)
return data
return wrapper
``` | non_test | qat callback doesn t work with multi input models or forward kwargs 🐛 bug the quantizationawaretraining can only handle models whose forward signature looks like forward data adding additional positional or keyword arguments breaks the forward propagation please reproduce using the boringmodel the boringmodel doesn t have the right forward signature to produce errors but i included an equivalently simple model below to reproduce python import functools import torch import torch nn as nn class simple nn module def init self super simple self init self conv nn kernel size self softmax nn softmax dim def forward self a b softmax false use positional arguments and keyword argument feat b self conv b feat a self conv a combined feat a feat b if softmax return self softmax combined else return combined property def example input array self return torch randn torch randn def wrap quantize forward context model func from quantizationawaretraining callback functools wraps func def wrapper data data model quant data data func data data model dequant data return data return wrapper works fine without quantization model simple input sample model example input array model input sample model input sample softmax true simulate the on fit start hook behavior model quant torch quantization quantstub model dequant torch quantization dequantstub model forward wrap quantize forward context model model forward model input sample model input sample softmax true the last two lines will raise the following exceptions typeerror wrapper takes positional argument but were given typeerror wrapper got an unexpected keyword argument softmax expected behavior forward prop should work fine for both the quantized and unquantized model environment cuda gpu geforce gtx with max q design available true version packages numpy pytorch debug false pytorch version pytorch lightning tqdm system os linux architecture processor python version ubuntu smp sat may utc additional context the solution to the minimal example is to make the decoration wrapper more generic python def wrap quantize forward context model func from quantizationawaretraining callback functools wraps func def wrapper data kwargs data model quant data data func data kwargs data model dequant data return data return wrapper however i notice that callbacks quantization py also includes a second decorator wrap qat forward context and i am not exactly clear on the intent behind having two separate decorators i see that wrap quantize forward context is only used in the on fit end hook any insight here borda i can make a pr for the above solution but first i tested it in an actual training by patching my local install python def wrap qat forward context quant cb model pl lightningmodule func callable trigger condition optional none callable functools wraps func def wrapper data kwargs any changes to decoration signature is func true isinstance trigger condition callable and trigger condition model trainer is count true isinstance trigger condition int and quant cb forward calls trigger condition quant run trigger condition is none or is func true or is count true if quant run quant cb forward calls data model quant data triggers attributeerror tuple object has no attribute detach data func data kwargs if quant run data model dequant data return data return wrapper so i tried also expanding the data before passing to model quant but this goes back to a typeerror python def wrap qat forward context quant cb model pl lightningmodule func callable trigger condition optional none callable functools wraps func def wrapper data kwargs any is func true isinstance trigger condition callable and trigger condition model trainer is count true isinstance trigger condition int and quant cb forward calls trigger condition quant run trigger condition is none or is func true or is count true if quant run quant cb forward calls data model quant data attempted tuple unpacking get typeerror forward takes positional arguments but were given data func data kwargs if quant run data model dequant data return data return wrapper | 0 |
37,495 | 5,117,350,862 | IssuesEvent | 2017-01-07 15:55:16 | d3athrow/vgstation13 | https://api.github.com/repos/d3athrow/vgstation13 | closed | 95% sure killer tomatoes have been broken since botany3 | Needs Moar Testing | In botany2, killer tomatoes were a random chance from tomato mutations, it went as follows
33% blue tomato
33% blood tomato
33% killer tomato
Now in botany3, killer tomato is a mutation of blood tomato. However ever since it's release I have yet to get a killer tomato outside of xenobotany means (random seed happens to have killer tomato sprite) and I've never seen them grown by other botanist either. I'm pretty sure that they're broken, as I have no problem with any other mutation except this very specific one.
| 1.0 | 95% sure killer tomatoes have been broken since botany3 - In botany2, killer tomatoes were a random chance from tomato mutations, it went as follows
33% blue tomato
33% blood tomato
33% killer tomato
Now in botany3, killer tomato is a mutation of blood tomato. However ever since it's release I have yet to get a killer tomato outside of xenobotany means (random seed happens to have killer tomato sprite) and I've never seen them grown by other botanist either. I'm pretty sure that they're broken, as I have no problem with any other mutation except this very specific one.
| test | sure killer tomatoes have been broken since in killer tomatoes were a random chance from tomato mutations it went as follows blue tomato blood tomato killer tomato now in killer tomato is a mutation of blood tomato however ever since it s release i have yet to get a killer tomato outside of xenobotany means random seed happens to have killer tomato sprite and i ve never seen them grown by other botanist either i m pretty sure that they re broken as i have no problem with any other mutation except this very specific one | 1 |
686,199 | 23,481,566,318 | IssuesEvent | 2022-08-17 11:03:03 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Crash returning to page with speedreader enabled | crash priority/P2 QA/Yes release-notes/exclude feature/speedreader OS/Desktop | IMPORTANT: Your crash has already been automatically reported to our crash system. Please file this bug only if you can provide more information about it.
Brave Version: 1.43.58 Chromium: 104.0.5112.81
Operating System: Linux 5.18.16-200.fc36.x86_64
URL (if applicable) where crash occurred: https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html
Can you reproduce this crash? No. Was reliable for a while, but currently not re-occuring.
What steps will reproduce this crash? (If it's not reproducible, what were you doing just before the crash?)
1. Visit https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html with speedreader enabled
2. Follow link to doc.rust-lang.org
3. Click back button to navigate back to the blog post
4. Tab reports segv
NB: Still occurs if I click 'Show original page' first.
****DO NOT CHANGE BELOW THIS LINE****
Crash ID: crash/1bfd0000-f3fb-a908-0000-000000000000 | 1.0 | Crash returning to page with speedreader enabled - IMPORTANT: Your crash has already been automatically reported to our crash system. Please file this bug only if you can provide more information about it.
Brave Version: 1.43.58 Chromium: 104.0.5112.81
Operating System: Linux 5.18.16-200.fc36.x86_64
URL (if applicable) where crash occurred: https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html
Can you reproduce this crash? No. Was reliable for a while, but currently not re-occuring.
What steps will reproduce this crash? (If it's not reproducible, what were you doing just before the crash?)
1. Visit https://blog.rust-lang.org/2022/08/11/Rust-1.63.0.html with speedreader enabled
2. Follow link to doc.rust-lang.org
3. Click back button to navigate back to the blog post
4. Tab reports segv
NB: Still occurs if I click 'Show original page' first.
****DO NOT CHANGE BELOW THIS LINE****
Crash ID: crash/1bfd0000-f3fb-a908-0000-000000000000 | non_test | crash returning to page with speedreader enabled important your crash has already been automatically reported to our crash system please file this bug only if you can provide more information about it brave version chromium operating system linux url if applicable where crash occurred can you reproduce this crash no was reliable for a while but currently not re occuring what steps will reproduce this crash if it s not reproducible what were you doing just before the crash visit with speedreader enabled follow link to doc rust lang org click back button to navigate back to the blog post tab reports segv nb still occurs if i click show original page first do not change below this line crash id crash | 0 |
198,261 | 14,970,502,961 | IssuesEvent | 2021-01-27 19:42:48 | istio/istio | https://api.github.com/repos/istio/istio | closed | Mitigate dockerhub image rate limitting | area/test and release | https://docs.docker.com/docker-hub/download-rate-limit/
CI Impact:
* Base image is on docker.io, pulled each run
* Kind image is on docker.io, pulled each run
* buildkit image is on docker.io, pulled each run
User impact:
* May be limited in number of images they can pull, especially proxy image | 1.0 | Mitigate dockerhub image rate limitting - https://docs.docker.com/docker-hub/download-rate-limit/
CI Impact:
* Base image is on docker.io, pulled each run
* Kind image is on docker.io, pulled each run
* buildkit image is on docker.io, pulled each run
User impact:
* May be limited in number of images they can pull, especially proxy image | test | mitigate dockerhub image rate limitting ci impact base image is on docker io pulled each run kind image is on docker io pulled each run buildkit image is on docker io pulled each run user impact may be limited in number of images they can pull especially proxy image | 1 |
432,962 | 12,500,673,408 | IssuesEvent | 2020-06-01 22:55:12 | ngscheurich/dice_magick | https://api.github.com/repos/ngscheurich/dice_magick | opened | Better test coverage | Priority: High Type: Maintenance | There's lots of untested code currently. We should try and get to > 90% test coverage and implement a test coverage reporter in CI, e.g. Coveralls. | 1.0 | Better test coverage - There's lots of untested code currently. We should try and get to > 90% test coverage and implement a test coverage reporter in CI, e.g. Coveralls. | non_test | better test coverage there s lots of untested code currently we should try and get to test coverage and implement a test coverage reporter in ci e g coveralls | 0 |
103,224 | 8,883,394,590 | IssuesEvent | 2019-01-14 15:35:45 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | Search: support people search via email on main search and also entity search control | fixed (test & close) noteworthy | for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by. | 1.0 | Search: support people search via email on main search and also entity search control - for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by. | test | search support people search via email on main search and also entity search control for cases where people have names with fewer characters and to speed along the search if someone has an email address to search by | 1 |
43,532 | 7,049,956,228 | IssuesEvent | 2018-01-03 01:48:45 | erdl/legacy-scrape-util | https://api.github.com/repos/erdl/legacy-scrape-util | closed | Create Notebook to obtain past egauge readings in the DHHL project | prio: high stat:in progress type: documentation type: new feature | Eileen pointed out readings are still missing for DHHL egauges. We were unsure how to execute Forrest script manually to retrieve past data, so I tried using the code **locally** to see how it could be done. [A page on the wiki was added documenting the process](https://github.com/erdl/scrape-util/wiki/Manually-Running-Scrape-Util-to-Retrieve-Past-Data). | 1.0 | Create Notebook to obtain past egauge readings in the DHHL project - Eileen pointed out readings are still missing for DHHL egauges. We were unsure how to execute Forrest script manually to retrieve past data, so I tried using the code **locally** to see how it could be done. [A page on the wiki was added documenting the process](https://github.com/erdl/scrape-util/wiki/Manually-Running-Scrape-Util-to-Retrieve-Past-Data). | non_test | create notebook to obtain past egauge readings in the dhhl project eileen pointed out readings are still missing for dhhl egauges we were unsure how to execute forrest script manually to retrieve past data so i tried using the code locally to see how it could be done | 0 |
486,617 | 14,011,910,872 | IssuesEvent | 2020-10-29 08:14:05 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.instagram.com - site is not usable | browser-fenix engine-gecko priority-critical | <!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/60643 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.instagram.com/
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201026092834</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/10/0294b4c0-26f7-46c6-abb4-63d73deebac2)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.instagram.com - site is not usable - <!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/60643 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.instagram.com/
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201026092834</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/10/0294b4c0-26f7-46c6-abb4-63d73deebac2)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
295,472 | 25,478,441,426 | IssuesEvent | 2022-11-25 16:56:57 | saleor/saleor-dashboard | https://api.github.com/repos/saleor/saleor-dashboard | opened | Cypress test fail: should recalculate weight after changing shipping weight unit. TC: SALEOR_0901 | tests | **Known bug for versions:**
v35: false
**Additional Info:**
Spec: As a staff user I want to change shop default weight unit | 1.0 | Cypress test fail: should recalculate weight after changing shipping weight unit. TC: SALEOR_0901 - **Known bug for versions:**
v35: false
**Additional Info:**
Spec: As a staff user I want to change shop default weight unit | test | cypress test fail should recalculate weight after changing shipping weight unit tc saleor known bug for versions false additional info spec as a staff user i want to change shop default weight unit | 1 |
57,013 | 6,536,298,751 | IssuesEvent | 2017-08-31 17:35:57 | w3c/web-platform-tests | https://api.github.com/repos/w3c/web-platform-tests | opened | Unhelpful failure reporting | testharness.js | Originally posted as https://github.com/w3c/testharness.js/issues/75 by @darobin on 27 May 2014, 14:09 UTC:
> In http://www.w3c-test.org/html/browsers/the-window-object/window-properties.html the first failure I get with Firefox, for test "EventTarget method: addEventListener", is:
>
> ```
> assert_equals: expected function "function addEventListener() { [native code] }" but got function "function addEventListener() { [native code] }"
> ```
| 1.0 | Unhelpful failure reporting - Originally posted as https://github.com/w3c/testharness.js/issues/75 by @darobin on 27 May 2014, 14:09 UTC:
> In http://www.w3c-test.org/html/browsers/the-window-object/window-properties.html the first failure I get with Firefox, for test "EventTarget method: addEventListener", is:
>
> ```
> assert_equals: expected function "function addEventListener() { [native code] }" but got function "function addEventListener() { [native code] }"
> ```
| test | unhelpful failure reporting originally posted as by darobin on may utc in the first failure i get with firefox for test eventtarget method addeventlistener is assert equals expected function function addeventlistener but got function function addeventlistener | 1 |
120,062 | 12,058,407,290 | IssuesEvent | 2020-04-15 17:24:19 | ift-gftc/doc.gdst | https://api.github.com/repos/ift-gftc/doc.gdst | closed | DOC | XML Mappings for Wild Harvest Events | documentation | ## 📖 Document
We need to add in XML Mappings for the Wild Harvest events. | 1.0 | DOC | XML Mappings for Wild Harvest Events - ## 📖 Document
We need to add in XML Mappings for the Wild Harvest events. | non_test | doc xml mappings for wild harvest events 📖 document we need to add in xml mappings for the wild harvest events | 0 |
260,122 | 22,593,347,116 | IssuesEvent | 2022-06-28 22:23:47 | phetsims/energy-skate-park | https://api.github.com/repos/phetsims/energy-skate-park | closed | CT Cannot read properties of undefined (reading 'typeName') | type:automated-testing | Also in basics:
```
energy-skate-park : fuzz : built-phet-io
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone
Query: fuzz&memoryLimit=1000&phetioStandalone
Uncaught TypeError: Cannot read properties of undefined (reading 'typeName')
TypeError: Cannot read properties of undefined (reading 'typeName')
at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576)
at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587)
at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939)
at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281)
at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239)
at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194)
at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846)
at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : fuzz : built-phet-io
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone
Query: fuzz&memoryLimit=1000&phetioStandalone
Uncaught TypeError: Cannot read properties of undefined (reading 'typeName')
TypeError: Cannot read properties of undefined (reading 'typeName')
at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576)
at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587)
at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939)
at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281)
at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239)
at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194)
at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846)
at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-fuzz : unbuilt
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655804863703%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-fuzz : unbuilt
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655819123377%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-wrappers-tests : assert
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park&phetioDebug=true&phetioWrapperDebug=true
27 out of 28 tests passed. 1 failed.
SimTests: energy-skate-park: iframe API failed:
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15)
at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32)
at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32
at Array.forEach (<anonymous>)
at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38)
at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32
at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9)
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-wrappers-tests : no-assert
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park
25 out of 26 tests passed. 1 failed.
SimTests: energy-skate-park: iframe API failed:
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15)
at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32)
at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32
at Array.forEach (<anonymous>)
at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38)
at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32
at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9)
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
``` | 1.0 | CT Cannot read properties of undefined (reading 'typeName') - Also in basics:
```
energy-skate-park : fuzz : built-phet-io
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone
Query: fuzz&memoryLimit=1000&phetioStandalone
Uncaught TypeError: Cannot read properties of undefined (reading 'typeName')
TypeError: Cannot read properties of undefined (reading 'typeName')
at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576)
at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587)
at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939)
at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281)
at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239)
at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194)
at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846)
at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655806752050%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : fuzz : built-phet-io
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone
Query: fuzz&memoryLimit=1000&phetioStandalone
Uncaught TypeError: Cannot read properties of undefined (reading 'typeName')
TypeError: Cannot read properties of undefined (reading 'typeName')
at h (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1031576)
at Object.getData (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14587)
at V.phetioStartEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1013939)
at V._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14472)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14180)
at V.set (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19281)
at V.set value [as value] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:19239)
at <unknown> (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:1962194)
at n.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:20846)
at m._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/build/phet-io/energy-skate-park_all_phet-io.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22fuzz%22%2C%22built-phet-io%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655820277967%7D&fuzz&memoryLimit=1000&phetioStandalone:952:14676)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-fuzz : unbuilt
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655804863703%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-fuzz : unbuilt
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/energy-skate-park/energy-skate-park_en.html?continuousTest=%7B%22test%22%3A%5B%22energy-skate-park%22%2C%22phet-io-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1655803244604%22%2C%22timestamp%22%3A1655819123377%7D&ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Query: ea&brand=phet-io&phetioStandalone&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-wrappers-tests : assert
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park&phetioDebug=true&phetioWrapperDebug=true
27 out of 28 tests passed. 1 failed.
SimTests: energy-skate-park: iframe API failed:
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15)
at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32)
at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32
at Array.forEach (<anonymous>)
at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38)
at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32
at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9)
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
----------------------------------
energy-skate-park : phet-io-wrappers-tests : no-assert
https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/phet-io-wrappers/phet-io-wrappers-tests.html?sim=energy-skate-park
25 out of 26 tests passed. 1 failed.
SimTests: energy-skate-park: iframe API failed:
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructor
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at Object.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:449:15)
at IOType.toStateObject (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/tandem/js/types/IOType.js:194:32)
at PhetioStateEngine.getValueJSON (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:107:49)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:165:32
at Array.forEach (<anonymous>)
at PhetioStateEngine.getState (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:161:38)
at PhetioStateEngine.initialize (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/PhetioStateEngine.js:439:30)
at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/phet-io/js/phetioEngine.js:238:32
at TinyProperty.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/chipper/dist/js/axon/js/TinyEmitter.js:68:9)
Uncaught Error: Uncaught Error: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
* mark the type as phetioState: false
* create a custom toStateObject method in your IO Type
* perhaps you have everything right, but forgot to pass in the IOType via phetioType in the constructorError: Assertion failed: fell back to root serialization state for energySkatePark.general.model.preferencesModel.skaterCharacterSetProperty. Potential solutions:
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1655803244604/assert/js/assert.js:28:13)
at assert (IOType.ts:531:14)
at toStateObject (IOType.ts:267:31)
at toStateObject (PhetioStateEngine.js:106:48)
at getValueJSON (PhetioStateEngine.js:165:33)
at Array.forEach
at forEach (PhetioStateEngine.js:161:37)
at getState (PhetioStateEngine.js:453:29)
at initialize (phetioEngine.js:250:31)
at listener (TinyEmitter.ts:94:8)
id: Bayes Puppeteer
Snapshot from 6/21/2022, 3:20:44 AM
``` | test | ct cannot read properties of undefined reading typename also in basics energy skate park fuzz built phet io query fuzz memorylimit phetiostandalone uncaught typeerror cannot read properties of undefined reading typename typeerror cannot read properties of undefined reading typename at h at object getdata at v phetiostartevent at v notifylisteners at v set at v set at v set value at at n emit at m notifylisteners id bayes puppeteer snapshot from am energy skate park fuzz built phet io query fuzz memorylimit phetiostandalone uncaught typeerror cannot read properties of undefined reading typename typeerror cannot read properties of undefined reading typename at h at object getdata at v phetiostartevent at v notifylisteners at v set at v set at v set value at at n emit at m notifylisteners id bayes puppeteer snapshot from am energy skate park phet io fuzz unbuilt query ea brand phet io phetiostandalone fuzz memorylimit uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io fuzz unbuilt query ea brand phet io phetiostandalone fuzz memorylimit uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io wrappers tests assert out of tests passed failed simtests energy skate park iframe api failed uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor at window assertions assertfunction at object tostateobject at iotype tostateobject at phetiostateengine getvaluejson at at array foreach at phetiostateengine getstate at phetiostateengine initialize at at tinyproperty emit uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructorerror assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am energy skate park phet io wrappers tests no assert out of tests passed failed simtests energy skate park iframe api failed uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructor at window assertions assertfunction at object tostateobject at iotype tostateobject at phetiostateengine getvaluejson at at array foreach at phetiostateengine getstate at phetiostateengine initialize at at tinyproperty emit uncaught error uncaught error assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions mark the type as phetiostate false create a custom tostateobject method in your io type perhaps you have everything right but forgot to pass in the iotype via phetiotype in the constructorerror assertion failed fell back to root serialization state for energyskatepark general model preferencesmodel skatercharactersetproperty potential solutions at window assertions assertfunction at assert iotype ts at tostateobject iotype ts at tostateobject phetiostateengine js at getvaluejson phetiostateengine js at array foreach at foreach phetiostateengine js at getstate phetiostateengine js at initialize phetioengine js at listener tinyemitter ts id bayes puppeteer snapshot from am | 1 |
55,103 | 6,426,548,207 | IssuesEvent | 2017-08-09 17:42:03 | tcpd/surf | https://api.github.com/repos/tcpd/surf | closed | Not able to read back a file generated by surf_worksheet_generator.R in Surf | Fixed. Waiting for Test P1 - High Priority | All the merges are lost. Needs to be fixed asap.
| 1.0 | Not able to read back a file generated by surf_worksheet_generator.R in Surf - All the merges are lost. Needs to be fixed asap.
| test | not able to read back a file generated by surf worksheet generator r in surf all the merges are lost needs to be fixed asap | 1 |
17,925 | 4,215,963,298 | IssuesEvent | 2016-06-30 07:24:18 | nemerosa/ontrack | https://api.github.com/repos/nemerosa/ontrack | opened | Generation of the Ontrack documentation | documentation | * [ ] Embed the Ontrack documentation in the code
* [ ] Generate both single page and PDF documentation
* [ ] Publish PDF documentation in GitHub release
* [ ] Publish HTML documentation in Ontrack GitHub pages
* [ ] Migrate the Wiki documentation to the embedded documentation
| 1.0 | Generation of the Ontrack documentation - * [ ] Embed the Ontrack documentation in the code
* [ ] Generate both single page and PDF documentation
* [ ] Publish PDF documentation in GitHub release
* [ ] Publish HTML documentation in Ontrack GitHub pages
* [ ] Migrate the Wiki documentation to the embedded documentation
| non_test | generation of the ontrack documentation embed the ontrack documentation in the code generate both single page and pdf documentation publish pdf documentation in github release publish html documentation in ontrack github pages migrate the wiki documentation to the embedded documentation | 0 |
33,281 | 7,694,311,749 | IssuesEvent | 2018-05-18 08:19:15 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | opened | JIT: assert when jitting Sse2.ConvertToInt32WithTruncation | area-CodeGen bug | ```C#
using System;
using System.Runtime.Intrinsics;
using System.Runtime.Intrinsics.X86;
class X
{
public static int Main()
{
Vector128<double> x = new Vector128<double>();
int y = Sse2.ConvertToInt32WithTruncation(x);
return y;
}
}
```
Running this with a CHECKED build hits an assert:
```
Assert failure(PID 25028 [0x000061c4], Thread: 14596 [0x3904]):
Assertion failed 'unreached' in 'X:Main():int' (IL size 23)
File: d:\repos\coreclr\src\jit\hwintrinsiccodegenxarch.cpp Line: 1103
Image: D:\repos\coreclr\bin\tests\Windows_NT.x64.Checked\Tests\Core_Root\CoreRun.exe
``` | 1.0 | JIT: assert when jitting Sse2.ConvertToInt32WithTruncation - ```C#
using System;
using System.Runtime.Intrinsics;
using System.Runtime.Intrinsics.X86;
class X
{
public static int Main()
{
Vector128<double> x = new Vector128<double>();
int y = Sse2.ConvertToInt32WithTruncation(x);
return y;
}
}
```
Running this with a CHECKED build hits an assert:
```
Assert failure(PID 25028 [0x000061c4], Thread: 14596 [0x3904]):
Assertion failed 'unreached' in 'X:Main():int' (IL size 23)
File: d:\repos\coreclr\src\jit\hwintrinsiccodegenxarch.cpp Line: 1103
Image: D:\repos\coreclr\bin\tests\Windows_NT.x64.Checked\Tests\Core_Root\CoreRun.exe
``` | non_test | jit assert when jitting c using system using system runtime intrinsics using system runtime intrinsics class x public static int main x new int y x return y running this with a checked build hits an assert assert failure pid thread assertion failed unreached in x main int il size file d repos coreclr src jit hwintrinsiccodegenxarch cpp line image d repos coreclr bin tests windows nt checked tests core root corerun exe | 0 |
333,664 | 29,801,340,309 | IssuesEvent | 2023-06-16 08:20:35 | bitsstdcheee/luogu-status | https://api.github.com/repos/bitsstdcheee/luogu-status | closed | 🛑 Luogu Contest is down | status luogu-contest | In [`74802cc`](https://github.com/bitsstdcheee/luogu-status/commit/74802cc4854ee533b23a229bcf3390051a6653c4
), Luogu Contest (https://www.luogu.com.cn/contest/list) was **down**:
- HTTP code: 403
- Response time: 12 ms
| 1.0 | 🛑 Luogu Contest is down - In [`74802cc`](https://github.com/bitsstdcheee/luogu-status/commit/74802cc4854ee533b23a229bcf3390051a6653c4
), Luogu Contest (https://www.luogu.com.cn/contest/list) was **down**:
- HTTP code: 403
- Response time: 12 ms
| test | 🛑 luogu contest is down in luogu contest was down http code response time ms | 1 |
31,830 | 26,170,631,530 | IssuesEvent | 2023-01-01 21:48:59 | ManimCommunity/manim | https://api.github.com/repos/ManimCommunity/manim | closed | Update Poetry Installer to New Version | infrastructure | The old poetry installer has been deprecated for a while now, but we still use it. In my short (but not perfect) testing, it appears to me that the old installer will have issues on 3.10 (Which as of the creation of this issue is 6 days from launch )
This will require large changes to our CI that I have struggled to do on my own, hence the issue. | 1.0 | Update Poetry Installer to New Version - The old poetry installer has been deprecated for a while now, but we still use it. In my short (but not perfect) testing, it appears to me that the old installer will have issues on 3.10 (Which as of the creation of this issue is 6 days from launch )
This will require large changes to our CI that I have struggled to do on my own, hence the issue. | non_test | update poetry installer to new version the old poetry installer has been deprecated for a while now but we still use it in my short but not perfect testing it appears to me that the old installer will have issues on which as of the creation of this issue is days from launch this will require large changes to our ci that i have struggled to do on my own hence the issue | 0 |
142,363 | 11,470,124,021 | IssuesEvent | 2020-02-09 00:16:46 | namhyung/uftrace | https://api.github.com/repos/namhyung/uftrace | opened | clang compiled uftrace test result | tests | I've compiled uftrace with clang-9.0.1 and run tests. The result is as follows:
```
Test case pg finstrument-fu
------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os
015 longjmp : OK NG NG NG NG OK OK OK OK OK
051 return : NG NG NG NG NG NG OK OK OK OK
052 nested_func : BI BI BI BI BI BI BI BI BI BI
062 arg_char : OK NG NG NG NG SK SK SK SK SK
082 arg_many : NG OK OK OK OK SK SK SK SK SK
083 arg_float : NG NG NG NG NG SK SK SK SK SK
084 arg_mixed : NG NG NG NG NG SK SK SK SK SK
085 arg_reg : NG NG NG NG NG SK SK SK SK SK
087 arg_variadic : NG NG NG NG NG SK SK SK SK SK
090 report_recursive : OK NG NG NG NG OK OK OK OK OK
124 exception : OK OK OK OK OK NG NG NG NG NG
125 report_range : OK NG OK OK OK OK OK OK OK OK
136 dynamic : BI BI BI BI BI BI BI BI BI BI
151 recv_runcmd : NG NG NG NG NG NG NG NG NG NG
162 pltbind_now_pie : BI BI BI BI BI BI BI BI BI BI
181 graph_full : NG NG NG NG NG NG NG NG NG NG
182 thread_exit : SG OK OK NG OK SG OK NG NG NG
184 arg_enum : NG NG NG NG NG NG NG NG NG NG
185 exception2 : OK OK OK OK OK NG NG NG NG NG
186 exception3 : OK OK OK OK OK NG NG NG NG NG
191 posix_spawn : NG OK OK OK OK NG OK OK OK OK
198 lib_arg_float : NG NG NG NG NG SK SK SK SK SK
201 arg_dwarf1 : NG NG NG NG NG SK SK SK SK SK
202 arg_dwarf2 : NG OK OK OK OK SK SK SK SK SK
203 arg_dwarf3 : NG NG NG NG NG SK SK SK SK SK
222 external_data : OK OK OK OK OK OK OK NG OK OK
225 dynamic_size : NG OK OK OK NG NG OK OK OK NG
228 read_pmu_cycle3 : OK NG NG NG NG OK OK NG OK OK
231 arg_bound : NG OK OK OK OK SK SK SK SK SK
232 dynamic_unpatch : BI BI BI BI BI BI BI BI BI BI
``` | 1.0 | clang compiled uftrace test result - I've compiled uftrace with clang-9.0.1 and run tests. The result is as follows:
```
Test case pg finstrument-fu
------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os
015 longjmp : OK NG NG NG NG OK OK OK OK OK
051 return : NG NG NG NG NG NG OK OK OK OK
052 nested_func : BI BI BI BI BI BI BI BI BI BI
062 arg_char : OK NG NG NG NG SK SK SK SK SK
082 arg_many : NG OK OK OK OK SK SK SK SK SK
083 arg_float : NG NG NG NG NG SK SK SK SK SK
084 arg_mixed : NG NG NG NG NG SK SK SK SK SK
085 arg_reg : NG NG NG NG NG SK SK SK SK SK
087 arg_variadic : NG NG NG NG NG SK SK SK SK SK
090 report_recursive : OK NG NG NG NG OK OK OK OK OK
124 exception : OK OK OK OK OK NG NG NG NG NG
125 report_range : OK NG OK OK OK OK OK OK OK OK
136 dynamic : BI BI BI BI BI BI BI BI BI BI
151 recv_runcmd : NG NG NG NG NG NG NG NG NG NG
162 pltbind_now_pie : BI BI BI BI BI BI BI BI BI BI
181 graph_full : NG NG NG NG NG NG NG NG NG NG
182 thread_exit : SG OK OK NG OK SG OK NG NG NG
184 arg_enum : NG NG NG NG NG NG NG NG NG NG
185 exception2 : OK OK OK OK OK NG NG NG NG NG
186 exception3 : OK OK OK OK OK NG NG NG NG NG
191 posix_spawn : NG OK OK OK OK NG OK OK OK OK
198 lib_arg_float : NG NG NG NG NG SK SK SK SK SK
201 arg_dwarf1 : NG NG NG NG NG SK SK SK SK SK
202 arg_dwarf2 : NG OK OK OK OK SK SK SK SK SK
203 arg_dwarf3 : NG NG NG NG NG SK SK SK SK SK
222 external_data : OK OK OK OK OK OK OK NG OK OK
225 dynamic_size : NG OK OK OK NG NG OK OK OK NG
228 read_pmu_cycle3 : OK NG NG NG NG OK OK NG OK OK
231 arg_bound : NG OK OK OK OK SK SK SK SK SK
232 dynamic_unpatch : BI BI BI BI BI BI BI BI BI BI
``` | test | clang compiled uftrace test result i ve compiled uftrace with clang and run tests the result is as follows test case pg finstrument fu os os longjmp ok ng ng ng ng ok ok ok ok ok return ng ng ng ng ng ng ok ok ok ok nested func bi bi bi bi bi bi bi bi bi bi arg char ok ng ng ng ng sk sk sk sk sk arg many ng ok ok ok ok sk sk sk sk sk arg float ng ng ng ng ng sk sk sk sk sk arg mixed ng ng ng ng ng sk sk sk sk sk arg reg ng ng ng ng ng sk sk sk sk sk arg variadic ng ng ng ng ng sk sk sk sk sk report recursive ok ng ng ng ng ok ok ok ok ok exception ok ok ok ok ok ng ng ng ng ng report range ok ng ok ok ok ok ok ok ok ok dynamic bi bi bi bi bi bi bi bi bi bi recv runcmd ng ng ng ng ng ng ng ng ng ng pltbind now pie bi bi bi bi bi bi bi bi bi bi graph full ng ng ng ng ng ng ng ng ng ng thread exit sg ok ok ng ok sg ok ng ng ng arg enum ng ng ng ng ng ng ng ng ng ng ok ok ok ok ok ng ng ng ng ng ok ok ok ok ok ng ng ng ng ng posix spawn ng ok ok ok ok ng ok ok ok ok lib arg float ng ng ng ng ng sk sk sk sk sk arg ng ng ng ng ng sk sk sk sk sk arg ng ok ok ok ok sk sk sk sk sk arg ng ng ng ng ng sk sk sk sk sk external data ok ok ok ok ok ok ok ng ok ok dynamic size ng ok ok ok ng ng ok ok ok ng read pmu ok ng ng ng ng ok ok ng ok ok arg bound ng ok ok ok ok sk sk sk sk sk dynamic unpatch bi bi bi bi bi bi bi bi bi bi | 1 |
636,120 | 20,592,545,103 | IssuesEvent | 2022-03-05 02:25:37 | apcountryman/picolibrary-microchip-megaavr | https://api.github.com/repos/apcountryman/picolibrary-microchip-megaavr | closed | Add Microchip megaAVR I2C controller interactive testing skeleton | priority-normal status-awaiting_review type-feature | Add Microchip megaAVR I2C controller (`::picolibrary::Microchip::megaAVR::Controller`) interactive testing skeleton.
- [x] Add `test/interactive/picolibrary/microchip/megaavr/i2c/controller/CMakeLists.txt`
- [x] Add the `test/interactive/picolibrary/microchip/megaavr/i2c/controller/` directory to the build | 1.0 | Add Microchip megaAVR I2C controller interactive testing skeleton - Add Microchip megaAVR I2C controller (`::picolibrary::Microchip::megaAVR::Controller`) interactive testing skeleton.
- [x] Add `test/interactive/picolibrary/microchip/megaavr/i2c/controller/CMakeLists.txt`
- [x] Add the `test/interactive/picolibrary/microchip/megaavr/i2c/controller/` directory to the build | non_test | add microchip megaavr controller interactive testing skeleton add microchip megaavr controller picolibrary microchip megaavr controller interactive testing skeleton add test interactive picolibrary microchip megaavr controller cmakelists txt add the test interactive picolibrary microchip megaavr controller directory to the build | 0 |
44,858 | 5,657,410,782 | IssuesEvent | 2017-04-10 07:00:20 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Add CI tests for geo-centroid | :Tilemap :Visualizations test v5.4.0 | the selenium tests should test the new geo-centroid feature, which is currently missing. | 1.0 | Add CI tests for geo-centroid - the selenium tests should test the new geo-centroid feature, which is currently missing. | test | add ci tests for geo centroid the selenium tests should test the new geo centroid feature which is currently missing | 1 |
274,488 | 23,842,521,910 | IssuesEvent | 2022-09-06 11:29:02 | stores-cedcommerce/Lucy-Store-Design | https://api.github.com/repos/stores-cedcommerce/Lucy-Store-Design | closed | The Refund policy title needed to be change. | Footer section Desktop Ready to test Content Type / typo fixed | **Actual result:**
The Refund policy title needed to be change.

**Expected result:**
The title needed to be updated like Returns and Refund Policy and the titles inside the page needed to be updated. | 1.0 | The Refund policy title needed to be change. - **Actual result:**
The Refund policy title needed to be change.

**Expected result:**
The title needed to be updated like Returns and Refund Policy and the titles inside the page needed to be updated. | test | the refund policy title needed to be change actual result the refund policy title needed to be change expected result the title needed to be updated like returns and refund policy and the titles inside the page needed to be updated | 1 |
348,079 | 31,465,781,990 | IssuesEvent | 2023-08-30 01:48:00 | hoprnet/hoprnet | https://api.github.com/repos/hoprnet/hoprnet | closed | loadtesting HOPRd nodes | bug testing stale | Related to https://github.com/Rpc-h/RPCh/pull/403
Running three types of load tests against RPCh + HOPRd nodes.
I've confirmed that the bottleneck is reproducible directly via using the HOPRd API so there isn't a bottleneck between RPCh <> HOPRd nodes (meaning the `rpc-server`).
Using two dedicated nodes, one acting as an entry and another as an exit.
image: `gcr.io/hoprassociation/hoprd:1682855128`
environment: `rpch staging`
The tests are meant to test `deliverability` and do not concern themselves with response time by intention.
Summary of the results of a single run:
- constant load test (run various load tests for 10 seconds each)
- 10 reqs in 10 seconds (1 req/s) = 100% delivery
- 100 reqs in 10 seconds (10 req/s) = 88% delivery
- 1000 reqs in 10 seconds (100 req/s) = 22% delivery
- burst load test (run various burst tests for 1 second each)
- 10 reqs in 1 second = 100% delivery
- 50 reqs in 1 second = 94% delivery
- 100 reqs in 1 second = 98% delivery
Additional context:
Why is this a problem for RPCh?
You can see that already with 10 reqs per second (which can easily be achieved by 1 RPCh user) we start experiencing large delivery losses. Additionally, as soon as the node's connected users are more than 1, the delivery losses keep increasing.
| 1.0 | loadtesting HOPRd nodes - Related to https://github.com/Rpc-h/RPCh/pull/403
Running three types of load tests against RPCh + HOPRd nodes.
I've confirmed that the bottleneck is reproducible directly via using the HOPRd API so there isn't a bottleneck between RPCh <> HOPRd nodes (meaning the `rpc-server`).
Using two dedicated nodes, one acting as an entry and another as an exit.
image: `gcr.io/hoprassociation/hoprd:1682855128`
environment: `rpch staging`
The tests are meant to test `deliverability` and do not concern themselves with response time by intention.
Summary of the results of a single run:
- constant load test (run various load tests for 10 seconds each)
- 10 reqs in 10 seconds (1 req/s) = 100% delivery
- 100 reqs in 10 seconds (10 req/s) = 88% delivery
- 1000 reqs in 10 seconds (100 req/s) = 22% delivery
- burst load test (run various burst tests for 1 second each)
- 10 reqs in 1 second = 100% delivery
- 50 reqs in 1 second = 94% delivery
- 100 reqs in 1 second = 98% delivery
Additional context:
Why is this a problem for RPCh?
You can see that already with 10 reqs per second (which can easily be achieved by 1 RPCh user) we start experiencing large delivery losses. Additionally, as soon as the node's connected users are more than 1, the delivery losses keep increasing.
| test | loadtesting hoprd nodes related to running three types of load tests against rpch hoprd nodes i ve confirmed that the bottleneck is reproducible directly via using the hoprd api so there isn t a bottleneck between rpch hoprd nodes meaning the rpc server using two dedicated nodes one acting as an entry and another as an exit image gcr io hoprassociation hoprd environment rpch staging the tests are meant to test deliverability and do not concern themselves with response time by intention summary of the results of a single run constant load test run various load tests for seconds each reqs in seconds req s delivery reqs in seconds req s delivery reqs in seconds req s delivery burst load test run various burst tests for second each reqs in second delivery reqs in second delivery reqs in second delivery additional context why is this a problem for rpch you can see that already with reqs per second which can easily be achieved by rpch user we start experiencing large delivery losses additionally as soon as the node s connected users are more than the delivery losses keep increasing | 1 |
340,482 | 30,518,856,204 | IssuesEvent | 2023-07-19 06:30:32 | guram-asanidze-1-btu-edu-ge/shualeduri | https://api.github.com/repos/guram-asanidze-1-btu-edu-ge/shualeduri | opened | 435d763 failed unit and formatting tests. | ci-pytest ci-black | Automatically generated message
435d7632962fde65708858b27fbea1ae9d0205b9 failed unit and formatting tests.
Pytest report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/pytest.html
Black report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/black.html
| 1.0 | 435d763 failed unit and formatting tests. - Automatically generated message
435d7632962fde65708858b27fbea1ae9d0205b9 failed unit and formatting tests.
Pytest report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/pytest.html
Black report: https://guram-asanidze-1-btu-edu-ge.github.io/shualeduri-ci/435d7632962fde65708858b27fbea1ae9d0205b9-1689748226/black.html
| test | failed unit and formatting tests automatically generated message failed unit and formatting tests pytest report black report | 1 |
223,479 | 17,602,475,289 | IssuesEvent | 2021-08-17 13:28:26 | Gemini-team/Gemini | https://api.github.com/repos/Gemini-team/Gemini | closed | Unit testing in Unity | enhancement testing | Until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended. This will not cut it going further, and the first step in the testing pipeline will be to get Unit testing in the Unity project up and running. | 1.0 | Unit testing in Unity - Until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended. This will not cut it going further, and the first step in the testing pipeline will be to get Unit testing in the Unity project up and running. | test | unit testing in unity until now there has only been manual testing of the code base by running the simulator and verifying that things works as intended this will not cut it going further and the first step in the testing pipeline will be to get unit testing in the unity project up and running | 1 |
25,693 | 4,165,930,802 | IssuesEvent | 2016-06-19 20:38:24 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Graphs only work if you are on first tab in Host and Container View | area/container area/host area/ui kind/bug status/resolved status/to-test | **Rancher Version:** master 3/16
**Steps to Reproduce:**
1. Go to Host View or Container View
2. Select Volume tab in container view or Port tab in host view
**Results:** Screen noticeably slow down and the graph stop showing data
**Expected:** show data
| 1.0 | Graphs only work if you are on first tab in Host and Container View - **Rancher Version:** master 3/16
**Steps to Reproduce:**
1. Go to Host View or Container View
2. Select Volume tab in container view or Port tab in host view
**Results:** Screen noticeably slow down and the graph stop showing data
**Expected:** show data
| test | graphs only work if you are on first tab in host and container view rancher version master steps to reproduce go to host view or container view select volume tab in container view or port tab in host view results screen noticeably slow down and the graph stop showing data expected show data | 1 |
317,827 | 27,270,490,381 | IssuesEvent | 2023-02-22 21:50:30 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | validate-modules for plugins: for callback plugins, type vs. callback_type | test support:core bug has_pr affects_2.15 | ##### SUMMARY
This is not a bug, but a discussion issue for one aspect of #71734 ([WIP] Extend validate-modules to also validate plugins). It provides details on one of the points in https://github.com/ansible/community/issues/560#issuecomment-691529907.
Callback plugins have a type. In the plugin class, it can be set as the class attribute `CALLBACK_TYPE`.
For documentation, there are two versions:
1. `callback_type: <type>`
https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/tree.py#L10
2. `type: <type>`
https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/default.py#L10
In ansible/ansible, `type` is much more common (there's only one callback using `callback_type`). In community.general, 1/3 of the callback plugins uses `callback_type`.
I would suggest that we use only one of these two. My personal suggestion would be `callback_type`, since it is close to the `CALLBACK_TYPE` attribute of the plugin class. On the other hand, @abadger chose `type` in the antsibull-docs schemas (https://github.com/ansible-community/antsibull/blob/main/antsibull/schemas/callback.py#L23).
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
test/lib/ansible_test/_data/sanity/validate-modules/validate_modules/main.py
##### ANSIBLE VERSION
2.11
bot_skip
| 1.0 | validate-modules for plugins: for callback plugins, type vs. callback_type - ##### SUMMARY
This is not a bug, but a discussion issue for one aspect of #71734 ([WIP] Extend validate-modules to also validate plugins). It provides details on one of the points in https://github.com/ansible/community/issues/560#issuecomment-691529907.
Callback plugins have a type. In the plugin class, it can be set as the class attribute `CALLBACK_TYPE`.
For documentation, there are two versions:
1. `callback_type: <type>`
https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/tree.py#L10
2. `type: <type>`
https://github.com/ansible/ansible/blob/d63d8803836df3231ba66f6e3a54576169d9f3a0/lib/ansible/plugins/callback/default.py#L10
In ansible/ansible, `type` is much more common (there's only one callback using `callback_type`). In community.general, 1/3 of the callback plugins uses `callback_type`.
I would suggest that we use only one of these two. My personal suggestion would be `callback_type`, since it is close to the `CALLBACK_TYPE` attribute of the plugin class. On the other hand, @abadger chose `type` in the antsibull-docs schemas (https://github.com/ansible-community/antsibull/blob/main/antsibull/schemas/callback.py#L23).
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
test/lib/ansible_test/_data/sanity/validate-modules/validate_modules/main.py
##### ANSIBLE VERSION
2.11
bot_skip
| test | validate modules for plugins for callback plugins type vs callback type summary this is not a bug but a discussion issue for one aspect of extend validate modules to also validate plugins it provides details on one of the points in callback plugins have a type in the plugin class it can be set as the class attribute callback type for documentation there are two versions callback type type in ansible ansible type is much more common there s only one callback using callback type in community general of the callback plugins uses callback type i would suggest that we use only one of these two my personal suggestion would be callback type since it is close to the callback type attribute of the plugin class on the other hand abadger chose type in the antsibull docs schemas issue type bug report component name test lib ansible test data sanity validate modules validate modules main py ansible version bot skip | 1 |
184,296 | 14,285,928,407 | IssuesEvent | 2020-11-23 14:32:46 | photoprism/photoprism | https://api.github.com/repos/photoprism/photoprism | closed | Physically moved files cannot be indexed again | bug important please-test | I'm encountering an issue with physically moving photos around where Photoprism thinks the moved photos are duplicates, even though they've been removed from their original position. My workflow is as follows:
* Copy all files from the SD card to an 'unsorted' directory
* Re-index so that they are accessible in Photoprism
* Edit them later and move them to the appropriate places
* Re-index the library
The problem is that in the 2nd reindexing: the photos appear to not yet have been removed from the database when it scans the new directory. It therefore tags the moved photos as duplicates. By the end of the re-indexing the photos are no longer found at their original locations, however they could not be scanned again either.
I've tried various methods including a complete re-scan of both the original directory and the new directory but PP doesn't recognise the new photo. My last resort would be to edit the database manually but it is not safe and inconvenient to do so every time..
The expectation is that Photoprism would be able to detect moved files with a single re-indexing. It seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed. I've also tried removing the sidecar completely from both directories and a clean re-indexing but that also didn't help.
This is an example log:
```
2020-10-25 08:27:32 INFO Indexing completed in 21 s
2020-10-25 08:27:31 INFO purge: searching index for hidden media files
2020-10-25 08:27:17 INFO index: skipped main raw file 2018/07/7R306756.ARW
2020-10-25 08:27:10 INFO index: no .ppignore file found
2020-10-25 08:27:10 INFO Indexing originals...
2020-10-25 08:27:10 INFO Settings saved
```
The RAW file, along with the JPG and XMP files are in the right location. No *.json files for them are present, however the json file is recreated after a normal re-indexing. It appears that it was pulled from the database and put back to the filesystem?.
To reproduce this issue:
* Add a file to PP by indexing
* Move the file elsewhere. The new directory might need to be named/placed in such a way that it is scanned before the old one in the directory hierarchy. In my case the old directory is a sub-directory of the new one.
* Re-index
As a workaround, the file could be reindexed if its entry is removed from the database in the "files" set. | 1.0 | Physically moved files cannot be indexed again - I'm encountering an issue with physically moving photos around where Photoprism thinks the moved photos are duplicates, even though they've been removed from their original position. My workflow is as follows:
* Copy all files from the SD card to an 'unsorted' directory
* Re-index so that they are accessible in Photoprism
* Edit them later and move them to the appropriate places
* Re-index the library
The problem is that in the 2nd reindexing: the photos appear to not yet have been removed from the database when it scans the new directory. It therefore tags the moved photos as duplicates. By the end of the re-indexing the photos are no longer found at their original locations, however they could not be scanned again either.
I've tried various methods including a complete re-scan of both the original directory and the new directory but PP doesn't recognise the new photo. My last resort would be to edit the database manually but it is not safe and inconvenient to do so every time..
The expectation is that Photoprism would be able to detect moved files with a single re-indexing. It seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed. I've also tried removing the sidecar completely from both directories and a clean re-indexing but that also didn't help.
This is an example log:
```
2020-10-25 08:27:32 INFO Indexing completed in 21 s
2020-10-25 08:27:31 INFO purge: searching index for hidden media files
2020-10-25 08:27:17 INFO index: skipped main raw file 2018/07/7R306756.ARW
2020-10-25 08:27:10 INFO index: no .ppignore file found
2020-10-25 08:27:10 INFO Indexing originals...
2020-10-25 08:27:10 INFO Settings saved
```
The RAW file, along with the JPG and XMP files are in the right location. No *.json files for them are present, however the json file is recreated after a normal re-indexing. It appears that it was pulled from the database and put back to the filesystem?.
To reproduce this issue:
* Add a file to PP by indexing
* Move the file elsewhere. The new directory might need to be named/placed in such a way that it is scanned before the old one in the directory hierarchy. In my case the old directory is a sub-directory of the new one.
* Re-index
As a workaround, the file could be reindexed if its entry is removed from the database in the "files" set. | test | physically moved files cannot be indexed again i m encountering an issue with physically moving photos around where photoprism thinks the moved photos are duplicates even though they ve been removed from their original position my workflow is as follows copy all files from the sd card to an unsorted directory re index so that they are accessible in photoprism edit them later and move them to the appropriate places re index the library the problem is that in the reindexing the photos appear to not yet have been removed from the database when it scans the new directory it therefore tags the moved photos as duplicates by the end of the re indexing the photos are no longer found at their original locations however they could not be scanned again either i ve tried various methods including a complete re scan of both the original directory and the new directory but pp doesn t recognise the new photo my last resort would be to edit the database manually but it is not safe and inconvenient to do so every time the expectation is that photoprism would be able to detect moved files with a single re indexing it seemed the sidecar files might be a reason for this as they are not automatically removed when the original files are removed i ve also tried removing the sidecar completely from both directories and a clean re indexing but that also didn t help this is an example log info indexing completed in s info purge searching index for hidden media files info index skipped main raw file arw info index no ppignore file found info indexing originals info settings saved the raw file along with the jpg and xmp files are in the right location no json files for them are present however the json file is recreated after a normal re indexing it appears that it was pulled from the database and put back to the filesystem to reproduce this issue add a file to pp by indexing move the file elsewhere the new directory might need to be named placed in such a way that it is scanned before the old one in the directory hierarchy in my case the old directory is a sub directory of the new one re index as a workaround the file could be reindexed if its entry is removed from the database in the files set | 1 |
80,581 | 10,193,503,894 | IssuesEvent | 2019-08-12 13:48:49 | catboost/catboost | https://api.github.com/repos/catboost/catboost | closed | wrong CatBoost Viewer installation instractions | bug documentation | Problem: {your page https://catboost.ai/docs/concepts/catboostviewer-installation.html#catboostviewer-installation sugests to click on npm but this references to broken link https://www.npmjs.com/package/nmp/tutorial. So how to install needed soft for CatBoost Viewer for use for Python , if possble at all}
catboost version: {latest}
Operating System: {Windows}
CPU: {i7}
# GPU: {GPU}
| 1.0 | wrong CatBoost Viewer installation instractions - Problem: {your page https://catboost.ai/docs/concepts/catboostviewer-installation.html#catboostviewer-installation sugests to click on npm but this references to broken link https://www.npmjs.com/package/nmp/tutorial. So how to install needed soft for CatBoost Viewer for use for Python , if possble at all}
catboost version: {latest}
Operating System: {Windows}
CPU: {i7}
# GPU: {GPU}
| non_test | wrong catboost viewer installation instractions problem your page sugests to click on npm but this references to broken link so how to install needed soft for catboost viewer for use for python if possble at all catboost version latest operating system windows cpu gpu gpu | 0 |
166,546 | 12,961,569,887 | IssuesEvent | 2020-07-20 15:51:20 | TBS-EACPD/infobase | https://api.github.com/repos/TBS-EACPD/infobase | closed | Add method of getting in touch to the report a problem widget | dead? enhancement user testing | There is no option in the Report a Problem Widget for users to receive an answer to potential inquiries they have about trouble with the site. There is an opportunity to use the InfoBase feedback inbox as a means of direct contact, once we reclaim access to that inbox. | 1.0 | Add method of getting in touch to the report a problem widget - There is no option in the Report a Problem Widget for users to receive an answer to potential inquiries they have about trouble with the site. There is an opportunity to use the InfoBase feedback inbox as a means of direct contact, once we reclaim access to that inbox. | test | add method of getting in touch to the report a problem widget there is no option in the report a problem widget for users to receive an answer to potential inquiries they have about trouble with the site there is an opportunity to use the infobase feedback inbox as a means of direct contact once we reclaim access to that inbox | 1 |
41,483 | 5,358,637,780 | IssuesEvent | 2017-02-20 23:01:17 | awslabs/s2n | https://api.github.com/repos/awslabs/s2n | closed | Add Test Integration with "BoGo" BoringSSL's Test Suite | type/test | BoringSSL has a test suite runner named "BoGo" that can run against other TLS Implementations which gets around 75% line coverage against BoringSSL. It is currently in an alpha stage for non-BoringSSL implementations, but it looks like they want to get it into a state where other implementations can use it too. I think it'd be interesting to get BoGo tests into a state where we can locally run BoGo through a "make bogo-test" or similar to see what, if any, BoGo tests s2n fails.
Once we have it in a state where we can run it manually, we can work on opting out of specific tests for features that s2n doesn't implement and work on bringing our pass rate up to 100%.
Links:
- https://www.ietf.org/mail-archive/web/tls/current/msg20793.html
- https://boringssl.googlesource.com/boringssl/+/master/ssl/test/
- https://boringssl.googlesource.com/boringssl/+/master/ssl/test/PORTING.md
- https://github.com/google/openssl-tests | 1.0 | Add Test Integration with "BoGo" BoringSSL's Test Suite - BoringSSL has a test suite runner named "BoGo" that can run against other TLS Implementations which gets around 75% line coverage against BoringSSL. It is currently in an alpha stage for non-BoringSSL implementations, but it looks like they want to get it into a state where other implementations can use it too. I think it'd be interesting to get BoGo tests into a state where we can locally run BoGo through a "make bogo-test" or similar to see what, if any, BoGo tests s2n fails.
Once we have it in a state where we can run it manually, we can work on opting out of specific tests for features that s2n doesn't implement and work on bringing our pass rate up to 100%.
Links:
- https://www.ietf.org/mail-archive/web/tls/current/msg20793.html
- https://boringssl.googlesource.com/boringssl/+/master/ssl/test/
- https://boringssl.googlesource.com/boringssl/+/master/ssl/test/PORTING.md
- https://github.com/google/openssl-tests | test | add test integration with bogo boringssl s test suite boringssl has a test suite runner named bogo that can run against other tls implementations which gets around line coverage against boringssl it is currently in an alpha stage for non boringssl implementations but it looks like they want to get it into a state where other implementations can use it too i think it d be interesting to get bogo tests into a state where we can locally run bogo through a make bogo test or similar to see what if any bogo tests fails once we have it in a state where we can run it manually we can work on opting out of specific tests for features that doesn t implement and work on bringing our pass rate up to links | 1 |
61,188 | 6,726,962,886 | IssuesEvent | 2017-10-17 11:56:17 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | app-linux-img-converter v1.2.3 (r3.2) | r3.2-fc26-testing | Update of app-linux-img-converter to v1.2.3 for Qubes r3.2, see comments below for details.
Built from: https://github.com/QubesOS/qubes-app-linux-img-converter/commit/851e4f026fcc9a45aa52e181360300640215064d
[Changes since previous version](https://github.com/QubesOS/qubes-app-linux-img-converter/compare/v1.2.2...v1.2.3):
QubesOS/qubes-app-linux-img-converter@851e4f0 version 1.2.3
QubesOS/qubes-app-linux-img-converter@004bc7b Removed .travis.yml bootstrap fix
QubesOS/qubes-app-linux-img-converter@1d757ce Updated manpage to have correct binary name
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current repo` (available 7 days from now)
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| 1.0 | app-linux-img-converter v1.2.3 (r3.2) - Update of app-linux-img-converter to v1.2.3 for Qubes r3.2, see comments below for details.
Built from: https://github.com/QubesOS/qubes-app-linux-img-converter/commit/851e4f026fcc9a45aa52e181360300640215064d
[Changes since previous version](https://github.com/QubesOS/qubes-app-linux-img-converter/compare/v1.2.2...v1.2.3):
QubesOS/qubes-app-linux-img-converter@851e4f0 version 1.2.3
QubesOS/qubes-app-linux-img-converter@004bc7b Removed .travis.yml bootstrap fix
QubesOS/qubes-app-linux-img-converter@1d757ce Updated manpage to have correct binary name
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current repo` (available 7 days from now)
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload app-linux-img-converter 851e4f026fcc9a45aa52e181360300640215064d r3.2 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| test | app linux img converter update of app linux img converter to for qubes see comments below for details built from qubesos qubes app linux img converter version qubesos qubes app linux img converter removed travis yml bootstrap fix qubesos qubes app linux img converter updated manpage to have correct binary name referenced issues if you re release manager you can issue gpg inline signed command upload app linux img converter current repo available days from now upload app linux img converter current dists repo you can choose subset of distributions like vm vm available days from now upload app linux img converter security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it | 1 |
612,882 | 19,058,387,479 | IssuesEvent | 2021-11-26 01:50:08 | vincetiu8/zombie-game | https://api.github.com/repos/vincetiu8/zombie-game | closed | Add particle hit effects | area/player area/enemies type/feature size/xs priority/low | Add particle effects to let the player know that they have hit a zombie or a zombie has hit a player. This can be number particles which tell how much damage was dealt, blood particles or both | 1.0 | Add particle hit effects - Add particle effects to let the player know that they have hit a zombie or a zombie has hit a player. This can be number particles which tell how much damage was dealt, blood particles or both | non_test | add particle hit effects add particle effects to let the player know that they have hit a zombie or a zombie has hit a player this can be number particles which tell how much damage was dealt blood particles or both | 0 |
170,694 | 13,198,368,989 | IssuesEvent | 2020-08-14 02:14:54 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | [FLAKEY TEST] end-to-end-geth-transfer-test -> celotool -> Transfer tests -> Transfers Frozen > -> light Node > -> when CeloGold is frozen -> when sender is whitelisted -> when sender is removed again from whitelist -> should not add the transaction to the pool | FLAKEY celotool end-to-end-geth-transfer-test | Discovered in commit 4ce1c02c6ad6ca55b90d889a680342fd43b7c221
Attempt No. 1:
AssertionError: expected 'Error: Transaction was not mined within 750 seconds, please make sure your transaction was properly sent. Be aware that it might still be mined!' to include 'Returned error: transfers are currently frozen'
at /home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:493:16
at Generator.throw (<anonymous>)
at rejected (/home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:8:65)
at process._tickCallback (internal/process/next_tick.js:68:7)
Attempt No. 2:
Test Passed!
| 1.0 | [FLAKEY TEST] end-to-end-geth-transfer-test -> celotool -> Transfer tests -> Transfers Frozen > -> light Node > -> when CeloGold is frozen -> when sender is whitelisted -> when sender is removed again from whitelist -> should not add the transaction to the pool - Discovered in commit 4ce1c02c6ad6ca55b90d889a680342fd43b7c221
Attempt No. 1:
AssertionError: expected 'Error: Transaction was not mined within 750 seconds, please make sure your transaction was properly sent. Be aware that it might still be mined!' to include 'Returned error: transfers are currently frozen'
at /home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:493:16
at Generator.throw (<anonymous>)
at rejected (/home/circleci/app/packages/celotool/src/e2e-tests/transfer_tests.ts:8:65)
at process._tickCallback (internal/process/next_tick.js:68:7)
Attempt No. 2:
Test Passed!
| test | end to end geth transfer test celotool transfer tests transfers frozen light node when celogold is frozen when sender is whitelisted when sender is removed again from whitelist should not add the transaction to the pool discovered in commit attempt no assertionerror expected error transaction was not mined within seconds please make sure your transaction was properly sent be aware that it might still be mined to include returned error transfers are currently frozen at home circleci app packages celotool src tests transfer tests ts at generator throw at rejected home circleci app packages celotool src tests transfer tests ts at process tickcallback internal process next tick js attempt no test passed | 1 |
175,779 | 13,609,877,886 | IssuesEvent | 2020-09-23 06:20:12 | nrwl/nx | https://api.github.com/repos/nrwl/nx | closed | Improve test performance | scope: testing tools type: feature | Hey there,
We've been using this package for quite some time now and it has improved the development experience at our company a lot!
However, as our project gets larger and larger, one thing that bothers us is the performance of unit tests. Our project is medium-sized with about 75 libraries and 3 apps (Angular and NestJS) and about 350 test suites.
It takes about 12 minutes on my machine (windows, very decent specs) to run all tests using `nx affected:test --all --parallel`. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. By using `nx affected`, we can run the unit tests for PRs in a few minutes, but for bigger PRs with more files, the time it takes to run the tests increases rapidly.
_NOTE: If this is something nobody else is experiencing, feel free to close this issue right away (please comment on how you achieved good test performance)._
After looking into a lot of issues, we're pretty confident that ts-jest and typescript compilation time are the main contributing factors. It takes between 10-20 seconds for each library just to get jest to start running the tests. Once jest is running, the tests are fast.
Using the fix from #1299 improved performance quite a bit (around 20-30%), but this still is slow for this amount of tests. kulshekhar/ts-jest#1115 is probably part of the reason.
We are still very confident that re-compiling all necessary files for each library individually is not a good approach for unit testing.
That's why we decided to try to run all our unit tests together. We altered the global `jest.config.js` and are now able to run all unit tests with the Jest CLI in a single test run.
The time it takes to run all those tests is about 60 seconds on my pc and 90-100 seconds on a laptop. I did not yet test the speed in our CI environment.
That's a massive speedup (over 10x faster), but could still be improved upon, and that's what this issue is actually about.
~~My suggestion is to alter `nx affected:test` to collect all the projects that have changed, then divide them by their needed jest setup (jest-preset-angular with JSDOM, just JSDOM or Node) and run those tests at once. This could be done by adding an entry to the `testMatch` array in the jest config dynamically for every library and would result in a maximum of three test runs.~~
EDIT:
After investigating a bit further, Jest supports a config entry called [`projects`](https://jestjs.io/docs/en/configuration.html#projects-arraystring--projectconfig) since Jest v20. This is a perfect fit for this problem.
If the global jest config was altered for every new project generated (just like `tsconfig.json` and `nx.json`), it would be possible to run jest from the root directly through the CLI. This would also allow configuring each project individually as needed through its own `jest.config.js` and to run `ng test <library>` just like now.
In addition, I would remove the [hack to alter the jest config in a script](https://github.com/nrwl/nx/blob/5109548710b090bf2fcd68596aed27c24fc45d97/packages/jest/src/builders/jest/jest.impl.ts#L69) and instead change the schematic that is used to generate a new angular library to add the config needed for 'jest-preset-angular' to the libraries' `jest.config.js`.
`nx affected:test` would then be even easier to implement then in my previous suggestion. Just replace the `projects` entry in the global jest config by an array of changed project paths.
Pros:
- Unit tests run **a lot** faster
- Jest config is cleaner and more exposed
- [Jest CLI](https://jestjs.io/docs/en/cli) can be used to run unit tests
- [VSCode-Jest](https://github.com/jest-community/vscode-jest) can be used to run unit tests
- The coverage report shows the total coverage
Cons:
- Parsing failed projects is probably a bit more difficult to implement
## Expected Behavior
Be able to run unit tests in 1-2 minutes for small and medium-sized projects and even less for PRs.
## Current Behavior
It takes about 12 minutes on my machine (windows, very decent specs) to run all tests. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes.
### Context
Please provide any relevant information about your setup:
@nrwl/angular : 8.11.0
@nrwl/cli : 8.11.0
@nrwl/cypress : 8.11.0
@nrwl/jest : 8.11.0
@nrwl/linter : 8.11.0
@nrwl/nest : 8.11.0
@nrwl/node : 8.11.0
@nrwl/tao : 8.11.0
@nrwl/workspace : 8.11.0
typescript : 3.4.5
jest-preset-angular: 7.1.1
jest: 24.8.0
ts-jest: 24.3.0 | 1.0 | Improve test performance - Hey there,
We've been using this package for quite some time now and it has improved the development experience at our company a lot!
However, as our project gets larger and larger, one thing that bothers us is the performance of unit tests. Our project is medium-sized with about 75 libraries and 3 apps (Angular and NestJS) and about 350 test suites.
It takes about 12 minutes on my machine (windows, very decent specs) to run all tests using `nx affected:test --all --parallel`. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes. By using `nx affected`, we can run the unit tests for PRs in a few minutes, but for bigger PRs with more files, the time it takes to run the tests increases rapidly.
_NOTE: If this is something nobody else is experiencing, feel free to close this issue right away (please comment on how you achieved good test performance)._
After looking into a lot of issues, we're pretty confident that ts-jest and typescript compilation time are the main contributing factors. It takes between 10-20 seconds for each library just to get jest to start running the tests. Once jest is running, the tests are fast.
Using the fix from #1299 improved performance quite a bit (around 20-30%), but this still is slow for this amount of tests. kulshekhar/ts-jest#1115 is probably part of the reason.
We are still very confident that re-compiling all necessary files for each library individually is not a good approach for unit testing.
That's why we decided to try to run all our unit tests together. We altered the global `jest.config.js` and are now able to run all unit tests with the Jest CLI in a single test run.
The time it takes to run all those tests is about 60 seconds on my pc and 90-100 seconds on a laptop. I did not yet test the speed in our CI environment.
That's a massive speedup (over 10x faster), but could still be improved upon, and that's what this issue is actually about.
~~My suggestion is to alter `nx affected:test` to collect all the projects that have changed, then divide them by their needed jest setup (jest-preset-angular with JSDOM, just JSDOM or Node) and run those tests at once. This could be done by adding an entry to the `testMatch` array in the jest config dynamically for every library and would result in a maximum of three test runs.~~
EDIT:
After investigating a bit further, Jest supports a config entry called [`projects`](https://jestjs.io/docs/en/configuration.html#projects-arraystring--projectconfig) since Jest v20. This is a perfect fit for this problem.
If the global jest config was altered for every new project generated (just like `tsconfig.json` and `nx.json`), it would be possible to run jest from the root directly through the CLI. This would also allow configuring each project individually as needed through its own `jest.config.js` and to run `ng test <library>` just like now.
In addition, I would remove the [hack to alter the jest config in a script](https://github.com/nrwl/nx/blob/5109548710b090bf2fcd68596aed27c24fc45d97/packages/jest/src/builders/jest/jest.impl.ts#L69) and instead change the schematic that is used to generate a new angular library to add the config needed for 'jest-preset-angular' to the libraries' `jest.config.js`.
`nx affected:test` would then be even easier to implement then in my previous suggestion. Just replace the `projects` entry in the global jest config by an array of changed project paths.
Pros:
- Unit tests run **a lot** faster
- Jest config is cleaner and more exposed
- [Jest CLI](https://jestjs.io/docs/en/cli) can be used to run unit tests
- [VSCode-Jest](https://github.com/jest-community/vscode-jest) can be used to run unit tests
- The coverage report shows the total coverage
Cons:
- Parsing failed projects is probably a bit more difficult to implement
## Expected Behavior
Be able to run unit tests in 1-2 minutes for small and medium-sized projects and even less for PRs.
## Current Behavior
It takes about 12 minutes on my machine (windows, very decent specs) to run all tests. In our CI environment (Linux) and on laptops this can take quite a bit longer, around 20 to 25 minutes.
### Context
Please provide any relevant information about your setup:
@nrwl/angular : 8.11.0
@nrwl/cli : 8.11.0
@nrwl/cypress : 8.11.0
@nrwl/jest : 8.11.0
@nrwl/linter : 8.11.0
@nrwl/nest : 8.11.0
@nrwl/node : 8.11.0
@nrwl/tao : 8.11.0
@nrwl/workspace : 8.11.0
typescript : 3.4.5
jest-preset-angular: 7.1.1
jest: 24.8.0
ts-jest: 24.3.0 | test | improve test performance hey there we ve been using this package for quite some time now and it has improved the development experience at our company a lot however as our project gets larger and larger one thing that bothers us is the performance of unit tests our project is medium sized with about libraries and apps angular and nestjs and about test suites it takes about minutes on my machine windows very decent specs to run all tests using nx affected test all parallel in our ci environment linux and on laptops this can take quite a bit longer around to minutes by using nx affected we can run the unit tests for prs in a few minutes but for bigger prs with more files the time it takes to run the tests increases rapidly note if this is something nobody else is experiencing feel free to close this issue right away please comment on how you achieved good test performance after looking into a lot of issues we re pretty confident that ts jest and typescript compilation time are the main contributing factors it takes between seconds for each library just to get jest to start running the tests once jest is running the tests are fast using the fix from improved performance quite a bit around but this still is slow for this amount of tests kulshekhar ts jest is probably part of the reason we are still very confident that re compiling all necessary files for each library individually is not a good approach for unit testing that s why we decided to try to run all our unit tests together we altered the global jest config js and are now able to run all unit tests with the jest cli in a single test run the time it takes to run all those tests is about seconds on my pc and seconds on a laptop i did not yet test the speed in our ci environment that s a massive speedup over faster but could still be improved upon and that s what this issue is actually about my suggestion is to alter nx affected test to collect all the projects that have changed then divide them by their needed jest setup jest preset angular with jsdom just jsdom or node and run those tests at once this could be done by adding an entry to the testmatch array in the jest config dynamically for every library and would result in a maximum of three test runs edit after investigating a bit further jest supports a config entry called since jest this is a perfect fit for this problem if the global jest config was altered for every new project generated just like tsconfig json and nx json it would be possible to run jest from the root directly through the cli this would also allow configuring each project individually as needed through its own jest config js and to run ng test just like now in addition i would remove the and instead change the schematic that is used to generate a new angular library to add the config needed for jest preset angular to the libraries jest config js nx affected test would then be even easier to implement then in my previous suggestion just replace the projects entry in the global jest config by an array of changed project paths pros unit tests run a lot faster jest config is cleaner and more exposed can be used to run unit tests can be used to run unit tests the coverage report shows the total coverage cons parsing failed projects is probably a bit more difficult to implement expected behavior be able to run unit tests in minutes for small and medium sized projects and even less for prs current behavior it takes about minutes on my machine windows very decent specs to run all tests in our ci environment linux and on laptops this can take quite a bit longer around to minutes context please provide any relevant information about your setup nrwl angular nrwl cli nrwl cypress nrwl jest nrwl linter nrwl nest nrwl node nrwl tao nrwl workspace typescript jest preset angular jest ts jest | 1 |
228,525 | 18,239,839,152 | IssuesEvent | 2021-10-01 11:33:32 | junit-team/junit5 | https://api.github.com/repos/junit-team/junit5 | opened | Support text blocks as files in `@CsvSource` | type: enhancement component: Jupiter theme: parameterized tests | ## Overview
PR #2721 introduced experimental support for _text blocks_ in `@CsvSource`; however, there is room for improvement.
A CSV line within a text block cannot currently contain a new line (`\n`), even if it is within a quoted string; whereas, this is supported when using `@CsvSource(value = ...)`.
Comments do not make sense in a single string in `@CsvSource(value = ...)`, but they do make sense within `@CsvSource(textBlock = ...)`.
## Deliverables
- [ ] Support new lines within "CSV lines" in text blocks
- [ ] Support comment lines beginning with `#` in text blocks
| 1.0 | Support text blocks as files in `@CsvSource` - ## Overview
PR #2721 introduced experimental support for _text blocks_ in `@CsvSource`; however, there is room for improvement.
A CSV line within a text block cannot currently contain a new line (`\n`), even if it is within a quoted string; whereas, this is supported when using `@CsvSource(value = ...)`.
Comments do not make sense in a single string in `@CsvSource(value = ...)`, but they do make sense within `@CsvSource(textBlock = ...)`.
## Deliverables
- [ ] Support new lines within "CSV lines" in text blocks
- [ ] Support comment lines beginning with `#` in text blocks
| test | support text blocks as files in csvsource overview pr introduced experimental support for text blocks in csvsource however there is room for improvement a csv line within a text block cannot currently contain a new line n even if it is within a quoted string whereas this is supported when using csvsource value comments do not make sense in a single string in csvsource value but they do make sense within csvsource textblock deliverables support new lines within csv lines in text blocks support comment lines beginning with in text blocks | 1 |
142,193 | 21,687,146,583 | IssuesEvent | 2022-05-09 12:24:45 | department-of-veterans-affairs/vets-design-system-documentation | https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation | closed | Flesh out Address block Component documentation | vsp-design-system-team va-address-block | ## This update is for:
- [ ] Content styleguide
- [x] Component
- [ ] Pattern
- [ ] Utility
- [ ] Other
## What is the name
Address block Component
## What is the nature of this update?
- [ ] How to build this component/pattern
- [x] When to use this component/pattern
- [x] When to use something else
- [x] Usage guidance
- [x] Accessibility
- [ ] Implementation
- [ ] Research insights
- [ ] Package information
## Additional Context
Currently there is no content in the design.va.gov [Address block Component page](https://design.va.gov/components/address-block) other than an embedded Storybook example of the component. We should apply the new Component template to Address block and flesh out this documentation, including links to the Content style guide.
We also need to add an example that shows a five-digit zip code, per the advice of Content. Designers have wondered if the full nine-digit zip code is a requirement (it's not) when they see the current Address block example in design.va.gov. | 1.0 | Flesh out Address block Component documentation - ## This update is for:
- [ ] Content styleguide
- [x] Component
- [ ] Pattern
- [ ] Utility
- [ ] Other
## What is the name
Address block Component
## What is the nature of this update?
- [ ] How to build this component/pattern
- [x] When to use this component/pattern
- [x] When to use something else
- [x] Usage guidance
- [x] Accessibility
- [ ] Implementation
- [ ] Research insights
- [ ] Package information
## Additional Context
Currently there is no content in the design.va.gov [Address block Component page](https://design.va.gov/components/address-block) other than an embedded Storybook example of the component. We should apply the new Component template to Address block and flesh out this documentation, including links to the Content style guide.
We also need to add an example that shows a five-digit zip code, per the advice of Content. Designers have wondered if the full nine-digit zip code is a requirement (it's not) when they see the current Address block example in design.va.gov. | non_test | flesh out address block component documentation this update is for content styleguide component pattern utility other what is the name address block component what is the nature of this update how to build this component pattern when to use this component pattern when to use something else usage guidance accessibility implementation research insights package information additional context currently there is no content in the design va gov other than an embedded storybook example of the component we should apply the new component template to address block and flesh out this documentation including links to the content style guide we also need to add an example that shows a five digit zip code per the advice of content designers have wondered if the full nine digit zip code is a requirement it s not when they see the current address block example in design va gov | 0 |
126,097 | 10,383,582,403 | IssuesEvent | 2019-09-10 09:57:40 | RIOT-OS/RIOT | https://api.github.com/repos/RIOT-OS/RIOT | closed | Testing | Area: CI Area: tests Community: help wanted State: stale | Automated unit tests with hardware in the loop (SAMR21 plugged on CI server?)
- Related issues
- [ ] #3363
- [ ] #3392
- [ ] #7871
- Related PRs
- [ ] #7653
- [x] #7845
- [x] #7906
Automated network functionality tests (e.g. RPL + UDP/PING tests through border router, multi-hop) in IoTLAB dev sites? leverage PiFleet more?
- Related issues
- [ ] #3252
On-board CI testing in IoT-LAB (as it will provide soon the possibility to add custom nodes)
- Related issues
- [ ]
General CI testing
- Related issues
- [ ] #2143
- [ ] #5319
- Related PRs
- [ ] #7258
- [ ] #7786 | 1.0 | Testing - Automated unit tests with hardware in the loop (SAMR21 plugged on CI server?)
- Related issues
- [ ] #3363
- [ ] #3392
- [ ] #7871
- Related PRs
- [ ] #7653
- [x] #7845
- [x] #7906
Automated network functionality tests (e.g. RPL + UDP/PING tests through border router, multi-hop) in IoTLAB dev sites? leverage PiFleet more?
- Related issues
- [ ] #3252
On-board CI testing in IoT-LAB (as it will provide soon the possibility to add custom nodes)
- Related issues
- [ ]
General CI testing
- Related issues
- [ ] #2143
- [ ] #5319
- Related PRs
- [ ] #7258
- [ ] #7786 | test | testing automated unit tests with hardware in the loop plugged on ci server related issues related prs automated network functionality tests e g rpl udp ping tests through border router multi hop in iotlab dev sites leverage pifleet more related issues on board ci testing in iot lab as it will provide soon the possibility to add custom nodes related issues general ci testing related issues related prs | 1 |
53,725 | 13,198,209,515 | IssuesEvent | 2020-08-14 01:42:18 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | GCC truncation warnings with --enable-debug on Fedora 32 | Type: Building good first issue |
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 32
Linux Kernel | 5.7.14-200.fc32.x86_64
Architecture | x86-64
ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6)
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings.
### Describe how to reproduce the problem
```
./autogen.sh && ./configure --enable-debug && make
```
### Include any warning/errors/backtraces from the system logs
```
CC os/linux/libzfs_mount_os.lo
CC os/linux/libzfs_pool_os.lo
CC os/linux/libzfs_sendrecv_os.lo
os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE);
| ~~~~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
297 | append_mntopt(MNTOPT_CONTEXT, prop,
| ~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1
make[3]: *** Waiting for unfinished jobs....
make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs'
make[2]: *** [Makefile:660: all-recursive] Error 1
make[2]: Leaving directory '/home/fedora/zfs/lib'
make[1]: *** [Makefile:870: all-recursive] Error 1
make[1]: Leaving directory '/home/fedora/zfs'
make: *** [Makefile:733: all] Error 2
``` | 1.0 | GCC truncation warnings with --enable-debug on Fedora 32 -
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 32
Linux Kernel | 5.7.14-200.fc32.x86_64
Architecture | x86-64
ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6)
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings.
### Describe how to reproduce the problem
```
./autogen.sh && ./configure --enable-debug && make
```
### Include any warning/errors/backtraces from the system logs
```
CC os/linux/libzfs_mount_os.lo
CC os/linux/libzfs_pool_os.lo
CC os/linux/libzfs_sendrecv_os.lo
os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE);
| ~~~~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
297 | append_mntopt(MNTOPT_CONTEXT, prop,
| ~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1
make[3]: *** Waiting for unfinished jobs....
make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs'
make[2]: *** [Makefile:660: all-recursive] Error 1
make[2]: Leaving directory '/home/fedora/zfs/lib'
make[1]: *** [Makefile:870: all-recursive] Error 1
make[1]: Leaving directory '/home/fedora/zfs'
make: *** [Makefile:733: all] Error 2
``` | non_test | gcc truncation warnings with enable debug on fedora system information type version name distribution name fedora distribution version linux kernel architecture zfs version master commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing i m seeing truncation warnings when building zfs using configure enable debug on fedora if i build without enable debug i don t see the warnings describe how to reproduce the problem autogen sh configure enable debug make include any warning errors backtraces from the system logs cc os linux libzfs mount os lo cc os linux libzfs pool os lo cc os linux libzfs sendrecv os lo os linux libzfs mount os c in function ‘zfs selinux setcontext’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt name context mntopts mtabopt b true os linux libzfs mount os c note ‘snprintf’ output or more bytes assuming into a destination of size snprintf tmp mnt line max quote s s s s name val os linux libzfs mount os c in function ‘zfs adjust mount options’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt mntopt context prop os linux libzfs mount os c note ‘snprintf’ output between and bytes into a destination of size snprintf tmp mnt line max quote s s s s name val all warnings being treated as errors make error make waiting for unfinished jobs make leaving directory home fedora zfs lib libzfs make error make leaving directory home fedora zfs lib make error make leaving directory home fedora zfs make error | 0 |
291,127 | 21,916,024,209 | IssuesEvent | 2022-05-21 20:54:29 | MonoGame/MonoGame | https://api.github.com/repos/MonoGame/MonoGame | opened | Upgrading the samples to 3.8.1 | Documentation Help Wanted Good First Issue | Throughout the years, MonoGame has accumulated a bunch of samples. Unfortunately they are all scattered across multiple repositories and are not all ready-to-use for MonoGame 3.8.1 which makes consuming the documentation complicated.
A list of those samples is [available here](https://docs.monogame.net/articles/samples.html).
We would to unify the samples under the same repository and have the sample documentation page be more straightforward.
There are 3 sub-tasks related to this issue.
## 1. Upgrading the official samples to MonoGame 3.8.1
This basically is upgrading [the current samples repository](https://github.com/MonoGame/MonoGame.Samples) to use MonoGame 3.8.1. This should be straightforward.
## 2. Upgrading the former XNA samples and merging them into the main samples repository
@SimonDarksideJ has [a giant repository archive of the old XNA 4.0 samples](https://github.com/SimonDarksideJ/XNAGameStudio) (including the ones from old websites now defunct). Some of them are now irrelevant, but it would be interesting to port the most relevant ones to MonoGame 3.8.1 and merge them into [the main samples](https://github.com/MonoGame/MonoGame.Samples) (Platformer and NeonShooter are already ported but not up-to-date with MG 3.8.1).
Be mindful of respecting the licenses of each of these samples.
We are allowed to use the Microsoft samples as long as we mention ```© Microsoft all rights reserved```.
## 3. Upgrading and merging more samples
@CartBlanche has even more [samples](https://github.com/CartBlanche/MonoGame-Samples) archived which we could pick, upgrade, and merge. They seem to be for macOS and some of them are networking samples (which we dropped), so there's some clean up to do. Most have precompiled xnb, so they may be unexploitable. Some seem to come from a book, so we have to be mindful of licenses there too. | 1.0 | Upgrading the samples to 3.8.1 - Throughout the years, MonoGame has accumulated a bunch of samples. Unfortunately they are all scattered across multiple repositories and are not all ready-to-use for MonoGame 3.8.1 which makes consuming the documentation complicated.
A list of those samples is [available here](https://docs.monogame.net/articles/samples.html).
We would to unify the samples under the same repository and have the sample documentation page be more straightforward.
There are 3 sub-tasks related to this issue.
## 1. Upgrading the official samples to MonoGame 3.8.1
This basically is upgrading [the current samples repository](https://github.com/MonoGame/MonoGame.Samples) to use MonoGame 3.8.1. This should be straightforward.
## 2. Upgrading the former XNA samples and merging them into the main samples repository
@SimonDarksideJ has [a giant repository archive of the old XNA 4.0 samples](https://github.com/SimonDarksideJ/XNAGameStudio) (including the ones from old websites now defunct). Some of them are now irrelevant, but it would be interesting to port the most relevant ones to MonoGame 3.8.1 and merge them into [the main samples](https://github.com/MonoGame/MonoGame.Samples) (Platformer and NeonShooter are already ported but not up-to-date with MG 3.8.1).
Be mindful of respecting the licenses of each of these samples.
We are allowed to use the Microsoft samples as long as we mention ```© Microsoft all rights reserved```.
## 3. Upgrading and merging more samples
@CartBlanche has even more [samples](https://github.com/CartBlanche/MonoGame-Samples) archived which we could pick, upgrade, and merge. They seem to be for macOS and some of them are networking samples (which we dropped), so there's some clean up to do. Most have precompiled xnb, so they may be unexploitable. Some seem to come from a book, so we have to be mindful of licenses there too. | non_test | upgrading the samples to throughout the years monogame has accumulated a bunch of samples unfortunately they are all scattered across multiple repositories and are not all ready to use for monogame which makes consuming the documentation complicated a list of those samples is we would to unify the samples under the same repository and have the sample documentation page be more straightforward there are sub tasks related to this issue upgrading the official samples to monogame this basically is upgrading to use monogame this should be straightforward upgrading the former xna samples and merging them into the main samples repository simondarksidej has including the ones from old websites now defunct some of them are now irrelevant but it would be interesting to port the most relevant ones to monogame and merge them into platformer and neonshooter are already ported but not up to date with mg be mindful of respecting the licenses of each of these samples we are allowed to use the microsoft samples as long as we mention © microsoft all rights reserved upgrading and merging more samples cartblanche has even more archived which we could pick upgrade and merge they seem to be for macos and some of them are networking samples which we dropped so there s some clean up to do most have precompiled xnb so they may be unexploitable some seem to come from a book so we have to be mindful of licenses there too | 0 |
182,617 | 30,874,397,501 | IssuesEvent | 2023-08-03 13:26:53 | elastic/elastic-charts | https://api.github.com/repos/elastic/elastic-charts | opened | [partition] improve pie chart labelling | enhancement :partition design design need | Even if I strongly believe in the goodness of the current implementation of labels in pie chart, I believe we need to make an effort to improve the followings:
- find a way to show every label (inside or outside). Right now there are situation where outside labels are not represented.
- by default don't rotate the labels inside the slices, if doesn't fit put the label outside
- improve alignment of wrapped lines inside the slices (don't make them start from different positions where is not necessary)
<img width="488" alt="Screenshot 2023-08-03 at 15 21 54" src="https://github.com/elastic/elastic-charts/assets/1421091/bbe4d658-9dab-4c53-a4fa-2ce038ff63c1">
| 2.0 | [partition] improve pie chart labelling - Even if I strongly believe in the goodness of the current implementation of labels in pie chart, I believe we need to make an effort to improve the followings:
- find a way to show every label (inside or outside). Right now there are situation where outside labels are not represented.
- by default don't rotate the labels inside the slices, if doesn't fit put the label outside
- improve alignment of wrapped lines inside the slices (don't make them start from different positions where is not necessary)
<img width="488" alt="Screenshot 2023-08-03 at 15 21 54" src="https://github.com/elastic/elastic-charts/assets/1421091/bbe4d658-9dab-4c53-a4fa-2ce038ff63c1">
| non_test | improve pie chart labelling even if i strongly believe in the goodness of the current implementation of labels in pie chart i believe we need to make an effort to improve the followings find a way to show every label inside or outside right now there are situation where outside labels are not represented by default don t rotate the labels inside the slices if doesn t fit put the label outside improve alignment of wrapped lines inside the slices don t make them start from different positions where is not necessary img width alt screenshot at src | 0 |
260,184 | 22,599,148,800 | IssuesEvent | 2022-06-29 07:33:11 | qaul/qaul.net | https://api.github.com/repos/qaul/qaul.net | closed | Testing Findings, January 14 2022 | Discussion Testing Tracking Issue | ## General
Due to a bug in the time calculation when sending RPC messages to the GUI, libqaul crashed (see logs below).
This crash was not transparent to the user, as the UI was still working. This led to confusion.
* [x] Fix libqaul bug time calculation bug
* libqaul crashed due to time.rs panic (duration became negative)
* Polkadot had the same problem: https://github.com/paritytech/polkadot/issues/4096
* [x] find safer way of using time calculations
* #295
## Better Testing Feedback
* [x] #291
* [x] Log libqaul within flutter and make it accessible for analyzation in futher tests.
* e.g. library quick_log https://pub.dev/packages/quick_log
* [x] Create the functionality to send the logs via email to the developers.
* [x] Create the functionality to send device soft/hardware information via email to the developers.
* [x] #297
* [x] #298
## UI Improvements
* [x] Android: Make System back button working
* [x] Mobile: Portrait view: Add gesture sliding functionality between the taps
* [x] Feed View: make user icons clickable
* [x] User Account View: show connectivity option in view
* Display IP addresses via which the user is reachable
* [x] #293
* [x] Display round trip time milliseconds on network view correctly.
## Further Development Suggestions
- Synchronize feed
- Make User Validation Interactive
- Automatically exchange further more private information
- Explain Idea behind it to the user
- Create anti-spam feature for feed messenger
----
## Logs
Libqaul Panic Log
```
flutter: sendRpc send 84 bytes
flutter: sendRpc success
thread '<unnamed>' panicked at 'overflow when subtracting durations', library/core/src/time.rs:940:31
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
flutter: sendRpc send 84 bytes
ERROR libqaul::rpc > "SendError(..)"
flutter: sendRpc success
```
| 1.0 | Testing Findings, January 14 2022 - ## General
Due to a bug in the time calculation when sending RPC messages to the GUI, libqaul crashed (see logs below).
This crash was not transparent to the user, as the UI was still working. This led to confusion.
* [x] Fix libqaul bug time calculation bug
* libqaul crashed due to time.rs panic (duration became negative)
* Polkadot had the same problem: https://github.com/paritytech/polkadot/issues/4096
* [x] find safer way of using time calculations
* #295
## Better Testing Feedback
* [x] #291
* [x] Log libqaul within flutter and make it accessible for analyzation in futher tests.
* e.g. library quick_log https://pub.dev/packages/quick_log
* [x] Create the functionality to send the logs via email to the developers.
* [x] Create the functionality to send device soft/hardware information via email to the developers.
* [x] #297
* [x] #298
## UI Improvements
* [x] Android: Make System back button working
* [x] Mobile: Portrait view: Add gesture sliding functionality between the taps
* [x] Feed View: make user icons clickable
* [x] User Account View: show connectivity option in view
* Display IP addresses via which the user is reachable
* [x] #293
* [x] Display round trip time milliseconds on network view correctly.
## Further Development Suggestions
- Synchronize feed
- Make User Validation Interactive
- Automatically exchange further more private information
- Explain Idea behind it to the user
- Create anti-spam feature for feed messenger
----
## Logs
Libqaul Panic Log
```
flutter: sendRpc send 84 bytes
flutter: sendRpc success
thread '<unnamed>' panicked at 'overflow when subtracting durations', library/core/src/time.rs:940:31
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
flutter: sendRpc send 84 bytes
ERROR libqaul::rpc > "SendError(..)"
flutter: sendRpc success
```
| test | testing findings january general due to a bug in the time calculation when sending rpc messages to the gui libqaul crashed see logs below this crash was not transparent to the user as the ui was still working this led to confusion fix libqaul bug time calculation bug libqaul crashed due to time rs panic duration became negative polkadot had the same problem find safer way of using time calculations better testing feedback log libqaul within flutter and make it accessible for analyzation in futher tests e g library quick log create the functionality to send the logs via email to the developers create the functionality to send device soft hardware information via email to the developers ui improvements android make system back button working mobile portrait view add gesture sliding functionality between the taps feed view make user icons clickable user account view show connectivity option in view display ip addresses via which the user is reachable display round trip time milliseconds on network view correctly further development suggestions synchronize feed make user validation interactive automatically exchange further more private information explain idea behind it to the user create anti spam feature for feed messenger logs libqaul panic log flutter sendrpc send bytes flutter sendrpc success thread panicked at overflow when subtracting durations library core src time rs note run with rust backtrace environment variable to display a backtrace flutter sendrpc send bytes error libqaul rpc senderror flutter sendrpc success | 1 |
227,487 | 18,065,158,449 | IssuesEvent | 2021-09-20 18:13:11 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts - discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component | failed-test test-cloud test-failure-flaky v7.15.0 Team:DataDiscovery | **Version: 7.15.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts**
**Stack Trace:**
```
Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="saved-query-management-save-changes-button"])
Wait timed out after 10053ms
at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9)
at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13)
at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12)
at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5)
at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5)
at SavedQueryManagementComponentService.updateCurrentlyLoadedQuery (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/saved_query_management_component.ts:76:5)
at Context.<anonymous> (test/functional/apps/discover/feature_controls/discover_security.ts:134:9)
at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
**Other test failures:**
- discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy
- discover feature controls discover feature controls security global discover read-only privileges "before all" hook for "shows discover navlink"
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2305/testReport/_ | 3.0 | [test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts - discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component - **Version: 7.15.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/discover/feature_controls/discover_security·ts**
**Stack Trace:**
```
Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="saved-query-management-save-changes-button"])
Wait timed out after 10053ms
at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9)
at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13)
at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12)
at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5)
at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5)
at SavedQueryManagementComponentService.updateCurrentlyLoadedQuery (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/saved_query_management_component.ts:76:5)
at Context.<anonymous> (test/functional/apps/discover/feature_controls/discover_security.ts:134:9)
at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
**Other test failures:**
- discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy
- discover feature controls discover feature controls security global discover read-only privileges "before all" hook for "shows discover navlink"
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2305/testReport/_ | test | chrome x pack ui functional x pack test functional apps discover feature controls discover security·ts discover feature controls discover feature controls security global discover all privileges allow saving changes to a currently loaded query via the saved query management component version class chrome x pack ui functional x pack test functional apps discover feature controls discover security·ts stack trace error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules selenium webdriver lib webdriver js at runmicrotasks at processticksandrejections internal process task queues js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at proxy clickbycssselector var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common find ts at testsubjects click var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common test subjects ts at savedquerymanagementcomponentservice updatecurrentlyloadedquery var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services saved query management component ts at context test functional apps discover feature controls discover security ts at object apply var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules kbn test target node functional test runner lib mocha wrap function js other test failures discover feature controls discover feature controls security global discover all privileges allow saving currently loaded query as a copy discover feature controls discover feature controls security global discover read only privileges before all hook for shows discover navlink test report | 1 |
390,505 | 26,864,277,577 | IssuesEvent | 2023-02-03 21:38:12 | mindsdb/lightwood | https://api.github.com/repos/mindsdb/lightwood | closed | Improve "ETSMixer" documentation | help wanted good first issue documentation hacktoberfest Community | This mixer currently inherits the docstring from the `sktime` mixer, which translates into duplicated docs :)
A simple, short description for the underlying forecasting algorithm would be good enough. | 1.0 | Improve "ETSMixer" documentation - This mixer currently inherits the docstring from the `sktime` mixer, which translates into duplicated docs :)
A simple, short description for the underlying forecasting algorithm would be good enough. | non_test | improve etsmixer documentation this mixer currently inherits the docstring from the sktime mixer which translates into duplicated docs a simple short description for the underlying forecasting algorithm would be good enough | 0 |
134,933 | 10,948,535,882 | IssuesEvent | 2019-11-26 09:05:57 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: jepsen/bank-multitable/parts-start-kill-2 failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/02d62674ad2f9ca16183184ea6552691506675f1
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=jepsen/bank-multitable/parts-start-kill-2 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1611136&tab=artifacts#/jepsen/bank-multitable/parts-start-kill-2
```
The test failed on branch=release-19.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191126-1611136/jepsen/bank-multitable/parts-start-kill-2/run_1
cluster.go:1783,jepsen.go:98,jepsen.go:138,jepsen.go:324,test_runner.go:697: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1611136-1574753130-40-n6cpu4:1-6 -- tar --transform s,^,cockroach/, -c -z -f cockroach.tgz cockroach returned:
stderr:
stdout:
teamcity-1611136-1574753130-40-n6cpu4: tar --transform s,^,cockroa...........
1:
2:
3:
4:
exit status 255
5:
6:
Error: exit status 255
: exit status 1
``` | 2.0 | roachtest: jepsen/bank-multitable/parts-start-kill-2 failed - SHA: https://github.com/cockroachdb/cockroach/commits/02d62674ad2f9ca16183184ea6552691506675f1
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=jepsen/bank-multitable/parts-start-kill-2 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1611136&tab=artifacts#/jepsen/bank-multitable/parts-start-kill-2
```
The test failed on branch=release-19.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20191126-1611136/jepsen/bank-multitable/parts-start-kill-2/run_1
cluster.go:1783,jepsen.go:98,jepsen.go:138,jepsen.go:324,test_runner.go:697: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1611136-1574753130-40-n6cpu4:1-6 -- tar --transform s,^,cockroach/, -c -z -f cockroach.tgz cockroach returned:
stderr:
stdout:
teamcity-1611136-1574753130-40-n6cpu4: tar --transform s,^,cockroa...........
1:
2:
3:
4:
exit status 255
5:
6:
Error: exit status 255
: exit status 1
``` | test | roachtest jepsen bank multitable parts start kill failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen bank multitable parts start kill pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts jepsen bank multitable parts start kill run cluster go jepsen go jepsen go jepsen go test runner go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity tar transform s cockroach c z f cockroach tgz cockroach returned stderr stdout teamcity tar transform s cockroa exit status error exit status exit status | 1 |
9,621 | 3,060,347,237 | IssuesEvent | 2015-08-14 20:08:22 | Esri/photo-survey | https://api.github.com/repos/Esri/photo-survey | closed | Photo Survey Application - Testing Survey Submission | help wanted test | Firefox Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Firefox Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
IE 10 & 11 Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
IE 10 & 11 Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Chrome Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Chrome Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Safari Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Safari Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
| 1.0 | Photo Survey Application - Testing Survey Submission - Firefox Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Firefox Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
IE 10 & 11 Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
IE 10 & 11 Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Chrome Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Chrome Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Safari Submission of Survey (Hosted Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
Safari Submission of Survey (On Premises (AGS) Feature Service)
- [x] Test with required questions (proper message displayed on submission)
- [x] Answer questions and make sure that the feature service is updated
- [x] Answer questions and make sure that the SURVEYOR field is updated in the feature service
- [x] Check profile to make survey it is updated to show # of surveys submitted
| test | photo survey application testing survey submission firefox submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted firefox submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted ie submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted ie submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted chrome submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted chrome submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted safari submission of survey hosted feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted safari submission of survey on premises ags feature service test with required questions proper message displayed on submission answer questions and make sure that the feature service is updated answer questions and make sure that the surveyor field is updated in the feature service check profile to make survey it is updated to show of surveys submitted | 1 |
170,335 | 13,184,070,480 | IssuesEvent | 2020-08-12 18:42:04 | microsoft/PowerToys | https://api.github.com/repos/microsoft/PowerToys | opened | Keyboard Manager requires test for the interaction between the keyboard hook and the UI | Area-Tests Product-Keyboard Shortcut Manager | We should add tests to ensure the correct state variables are set when opening the Remap keys/shortcut window and when clicking the Type button, since the state variables affect the logic of the hook (for eg. it suppresses all key events while the Type window is in the foreground). This should also test that the key remaps are active while the Remap Shortcuts window is in the foreground.
The tests will involve testing the read/writes of the `KeyboardManagerUIState` variable (https://github.com/microsoft/PowerToys/blob/master/src/modules/keyboardmanager/common/KeyboardManagerState.h#L45)
| 1.0 | Keyboard Manager requires test for the interaction between the keyboard hook and the UI - We should add tests to ensure the correct state variables are set when opening the Remap keys/shortcut window and when clicking the Type button, since the state variables affect the logic of the hook (for eg. it suppresses all key events while the Type window is in the foreground). This should also test that the key remaps are active while the Remap Shortcuts window is in the foreground.
The tests will involve testing the read/writes of the `KeyboardManagerUIState` variable (https://github.com/microsoft/PowerToys/blob/master/src/modules/keyboardmanager/common/KeyboardManagerState.h#L45)
| test | keyboard manager requires test for the interaction between the keyboard hook and the ui we should add tests to ensure the correct state variables are set when opening the remap keys shortcut window and when clicking the type button since the state variables affect the logic of the hook for eg it suppresses all key events while the type window is in the foreground this should also test that the key remaps are active while the remap shortcuts window is in the foreground the tests will involve testing the read writes of the keyboardmanageruistate variable | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.