Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
214,428
16,588,077,081
IssuesEvent
2021-06-01 01:59:37
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Pessoa Desenvolvedora Back-end (Elixir) @ Magnetis Investimentos
CLT Docker Git Pleno Remoto Scrum Stale TDD Testes automatizados
## Nossa empresa Cada pessoa desenvolvedora nos ajuda a construir um produto cada vez melhor, participando do ciclo de concepção de novas funcionalidades, auxiliando nas decisões técnicas e realizando manutenção do código produzido. Nos organizamos em times multidisciplinares, que permitem que todos participem da concepção, planejamento, testes e lançamento do produto. Entendemos que o segredo do sucesso é o compartilhamento de conhecimento, comunicação e pair programming uma vez que todo nosso time está remoto em diversos estados do Brasil. Nós valorizamos pessoas que reconhecem as próprias falhas e falam abertamente sobre elas. Errar não é um tabu dentro do time e esperamos um plano de ação dos aprendizados. Autonomia e responsabilidade são outros dois pontos vitais que fazem com que as entregas sejam cumpridas em dia sem microgerenciamento. ## Descrição da vaga Estamos buscando uma pessoa que tenha interesse em trabalhar em ambiente de desenvolvimento de produtos, com autonomia e com mudanças constantes. ## Local Trabalho 100% remoto ## Requisitos **Obrigatórios:** - Experiência com desenvolvimento em Elixir (ou linguagem correlata como Node.js e Clojure.js) para aplicações críticas ao negócio. - Experiência com desenvolvimento de software voltado para qualidade (TDD ou BDD). - Conhecimento em versionamento de código utilizando Git/Github/GitLab. - Experiência com equipes ágeis (Kanban, scrum,XP). - Habilidade de atuar no processo de definição de arquitetura de soluções de software. - Conhecimento em inglês para escrita de comentários, revisão de PRs e leitura. **Desejáveis:** - Experiência com escrita de testes automatizados. - Habilidade em design de aplicações pensando em boas práticas de integração, performance e segurança. - Experiência com mentoria de novos membros do time. - Experiência com em Docker/ Google Cloud. **Diferenciais:** - Transparência. - Proatividade. - Paixão por desenvolvimento de produto e/ ou mercado financeiro. - Ótima comunicação escrita e oral. - Comprometimento com prazos e entregas. ## Benefícios - Assistência médica e odontológica (sem coparticipação); - Vale-refeição/ Vale-alimentação; - Verba Home Office; - Gympass - Oportunidade de ganhar Stock Option (participação na empresa); - Possibilidade de investir na plataforma a partir de R$1 sem pagar pela taxa de serviço; - Verba educacional para participação em cursos e eventos; - Trabalho Remoto. **Diferenciais:** - Autogestão do tempo ## Contratação CLT ## Como se candidatar Link para se cadastrar: https://jobs.kenoby.com/magnetis/job/pessoa-desenvolvedora-backend-elixir/605e244a7fe2e26df9603fbb?utm_source=website ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 07 dias após cada processo. E-mail para contato em caso de não haver resposta: felipe.silveira@magnetis.com.br ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Remoto #### Regime - CLT #### Nível - Pleno - Sênior
1.0
[Remoto] Pessoa Desenvolvedora Back-end (Elixir) @ Magnetis Investimentos - ## Nossa empresa Cada pessoa desenvolvedora nos ajuda a construir um produto cada vez melhor, participando do ciclo de concepção de novas funcionalidades, auxiliando nas decisões técnicas e realizando manutenção do código produzido. Nos organizamos em times multidisciplinares, que permitem que todos participem da concepção, planejamento, testes e lançamento do produto. Entendemos que o segredo do sucesso é o compartilhamento de conhecimento, comunicação e pair programming uma vez que todo nosso time está remoto em diversos estados do Brasil. Nós valorizamos pessoas que reconhecem as próprias falhas e falam abertamente sobre elas. Errar não é um tabu dentro do time e esperamos um plano de ação dos aprendizados. Autonomia e responsabilidade são outros dois pontos vitais que fazem com que as entregas sejam cumpridas em dia sem microgerenciamento. ## Descrição da vaga Estamos buscando uma pessoa que tenha interesse em trabalhar em ambiente de desenvolvimento de produtos, com autonomia e com mudanças constantes. ## Local Trabalho 100% remoto ## Requisitos **Obrigatórios:** - Experiência com desenvolvimento em Elixir (ou linguagem correlata como Node.js e Clojure.js) para aplicações críticas ao negócio. - Experiência com desenvolvimento de software voltado para qualidade (TDD ou BDD). - Conhecimento em versionamento de código utilizando Git/Github/GitLab. - Experiência com equipes ágeis (Kanban, scrum,XP). - Habilidade de atuar no processo de definição de arquitetura de soluções de software. - Conhecimento em inglês para escrita de comentários, revisão de PRs e leitura. **Desejáveis:** - Experiência com escrita de testes automatizados. - Habilidade em design de aplicações pensando em boas práticas de integração, performance e segurança. - Experiência com mentoria de novos membros do time. - Experiência com em Docker/ Google Cloud. **Diferenciais:** - Transparência. - Proatividade. - Paixão por desenvolvimento de produto e/ ou mercado financeiro. - Ótima comunicação escrita e oral. - Comprometimento com prazos e entregas. ## Benefícios - Assistência médica e odontológica (sem coparticipação); - Vale-refeição/ Vale-alimentação; - Verba Home Office; - Gympass - Oportunidade de ganhar Stock Option (participação na empresa); - Possibilidade de investir na plataforma a partir de R$1 sem pagar pela taxa de serviço; - Verba educacional para participação em cursos e eventos; - Trabalho Remoto. **Diferenciais:** - Autogestão do tempo ## Contratação CLT ## Como se candidatar Link para se cadastrar: https://jobs.kenoby.com/magnetis/job/pessoa-desenvolvedora-backend-elixir/605e244a7fe2e26df9603fbb?utm_source=website ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 07 dias após cada processo. E-mail para contato em caso de não haver resposta: felipe.silveira@magnetis.com.br ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Remoto #### Regime - CLT #### Nível - Pleno - Sênior
test
pessoa desenvolvedora back end elixir magnetis investimentos nossa empresa cada pessoa desenvolvedora nos ajuda a construir um produto cada vez melhor participando do ciclo de concepção de novas funcionalidades auxiliando nas decisões técnicas e realizando manutenção do código produzido nos organizamos em times multidisciplinares que permitem que todos participem da concepção planejamento testes e lançamento do produto entendemos que o segredo do sucesso é o compartilhamento de conhecimento comunicação e pair programming uma vez que todo nosso time está remoto em diversos estados do brasil nós valorizamos pessoas que reconhecem as próprias falhas e falam abertamente sobre elas errar não é um tabu dentro do time e esperamos um plano de ação dos aprendizados autonomia e responsabilidade são outros dois pontos vitais que fazem com que as entregas sejam cumpridas em dia sem microgerenciamento descrição da vaga estamos buscando uma pessoa que tenha interesse em trabalhar em ambiente de desenvolvimento de produtos com autonomia e com mudanças constantes local trabalho remoto requisitos obrigatórios experiência com desenvolvimento em elixir ou linguagem correlata como node js e clojure js para aplicações críticas ao negócio experiência com desenvolvimento de software voltado para qualidade tdd ou bdd conhecimento em versionamento de código utilizando git github gitlab experiência com equipes ágeis kanban scrum xp habilidade de atuar no processo de definição de arquitetura de soluções de software conhecimento em inglês para escrita de comentários revisão de prs e leitura desejáveis experiência com escrita de testes automatizados habilidade em design de aplicações pensando em boas práticas de integração performance e segurança experiência com mentoria de novos membros do time experiência com em docker google cloud diferenciais transparência proatividade paixão por desenvolvimento de produto e ou mercado financeiro ótima comunicação escrita e oral comprometimento com prazos e entregas benefícios assistência médica e odontológica sem coparticipação vale refeição vale alimentação verba home office gympass oportunidade de ganhar stock option participação na empresa possibilidade de investir na plataforma a partir de r sem pagar pela taxa de serviço verba educacional para participação em cursos e eventos trabalho remoto diferenciais autogestão do tempo contratação clt como se candidatar link para se cadastrar tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta felipe silveira magnetis com br labels alocação remoto regime clt nível pleno sênior
1
193,104
15,367,281,091
IssuesEvent
2021-03-02 02:57:11
taghead/taghead.xyz
https://api.github.com/repos/taghead/taghead.xyz
opened
Documentation issues with dependencies and setting up
documentation
[README.md](/README.md) section [1.4. Setting up Development Environment](https://github.com/taghead/taghead.xyz/blob/main/README.md#14-setting-up-development-environment). `npm install --only=dev` needs to be changed to `npm install` for the environment to work. [README.md](/README.md) section [1.1. Dependencies](https://github.com/taghead/taghead.xyz/blob/main/README.md#11-dependencies) should also contain `npm install -g typescript` for stability.
1.0
Documentation issues with dependencies and setting up - [README.md](/README.md) section [1.4. Setting up Development Environment](https://github.com/taghead/taghead.xyz/blob/main/README.md#14-setting-up-development-environment). `npm install --only=dev` needs to be changed to `npm install` for the environment to work. [README.md](/README.md) section [1.1. Dependencies](https://github.com/taghead/taghead.xyz/blob/main/README.md#11-dependencies) should also contain `npm install -g typescript` for stability.
non_test
documentation issues with dependencies and setting up readme md section npm install only dev needs to be changed to npm install for the environment to work readme md section should also contain npm install g typescript for stability
0
305,090
26,360,438,299
IssuesEvent
2023-01-11 13:01:07
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix methods.test_numpy_matrix_property_A
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3871055481/jobs/6598494206" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3858660206/jobs/6577425783" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3870274071/jobs/6597072064" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3860079109/jobs/6580170344" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-09T06:10:34.8500245Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-09T06:10:34.8500653Z E 2023-01-09T06:10:34.8501387Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAoAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-09T03:10:15.5788078Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-09T03:10:15.5788394Z E 2023-01-09T03:10:15.5789032Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAkAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-07T03:07:57.5741172Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-07T03:07:57.5741421Z E 2023-01-07T03:07:57.5741920Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAoAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details>
1.0
Fix methods.test_numpy_matrix_property_A - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3871055481/jobs/6598494206" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3858660206/jobs/6577425783" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3870274071/jobs/6597072064" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3860079109/jobs/6580170344" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-09T06:10:34.8500245Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-09T06:10:34.8500653Z E 2023-01-09T06:10:34.8501387Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAoAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-09T03:10:15.5788078Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-09T03:10:15.5788394Z E 2023-01-09T03:10:15.5789032Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAkAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_matrix/test_methods.py::test_numpy_matrix_property_A[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-07T03:07:57.5741172Z E ValueError: invalid literal for int() with base 10: '-1e-05' 2023-01-07T03:07:57.5741421Z E 2023-01-07T03:07:57.5741920Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AAoAAAAAAAAAAcEAAAABGtHy') as a decorator on your test case </details>
test
fix methods test numpy matrix property a tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test numpy test matrix test methods py test numpy matrix property a e valueerror invalid literal for int with base e e you can reproduce this example by temporarily adding reproduce failure b aaoaaaaaaaaaaceaaaabgthy as a decorator on your test case failed ivy tests test ivy test frontends test numpy test matrix test methods py test numpy matrix property a e valueerror invalid literal for int with base e e you can reproduce this example by temporarily adding reproduce failure b aakaaaaaaaaaaceaaaabgthy as a decorator on your test case failed ivy tests test ivy test frontends test numpy test matrix test methods py test numpy matrix property a e valueerror invalid literal for int with base e e you can reproduce this example by temporarily adding reproduce failure b aaoaaaaaaaaaaceaaaabgthy as a decorator on your test case
1
201,097
7,022,416,466
IssuesEvent
2017-12-22 10:28:02
JKGDevs/JediKnightGalaxies
https://api.github.com/repos/JKGDevs/JediKnightGalaxies
closed
Taunt Breaks Animations
bug priority:medium
After using /taunt your animation gets stuck and you no longer have a visible weapon (from first person) and you are unable to reload. You can still fire the weapon until the clip runs out. You can also no longer taunt. This happens until you respawn.
1.0
Taunt Breaks Animations - After using /taunt your animation gets stuck and you no longer have a visible weapon (from first person) and you are unable to reload. You can still fire the weapon until the clip runs out. You can also no longer taunt. This happens until you respawn.
non_test
taunt breaks animations after using taunt your animation gets stuck and you no longer have a visible weapon from first person and you are unable to reload you can still fire the weapon until the clip runs out you can also no longer taunt this happens until you respawn
0
149,431
11,900,775,676
IssuesEvent
2020-03-30 11:16:12
astpl1998/Kanam-Latex
https://api.github.com/repos/astpl1998/Kanam-Latex
closed
RMG-Email trigger for Sales Invoice authorization_PFS
17.Testing2_Completed
Dear Team, Here am created new issue for "Email trigger for Sales Invoice authorization development". In future i have added PFS for further process. Thanks and Regards, S.Dhanwandher.
1.0
RMG-Email trigger for Sales Invoice authorization_PFS - Dear Team, Here am created new issue for "Email trigger for Sales Invoice authorization development". In future i have added PFS for further process. Thanks and Regards, S.Dhanwandher.
test
rmg email trigger for sales invoice authorization pfs dear team here am created new issue for email trigger for sales invoice authorization development in future i have added pfs for further process thanks and regards s dhanwandher
1
438,326
30,636,805,471
IssuesEvent
2023-07-24 18:25:48
cal-itp/customer-success
https://api.github.com/repos/cal-itp/customer-success
closed
Develop Customer Success Philosophy
documentation customer-resource
- [x] Information gather current outreach procedures and initiatives with @o-ram - [ ] Document functional duties for outreach #290 - [x] #247 - [x] Research best approaches to customer success philosophy in praxis - [x] Draft customer success philosophy - [x] Draft review by @o-ram @AnthonyRollins @laneymangan Jess and any other relevant parties - [x] Edit and provide second draft - [x] Second draft review - [x] Submit draft for Cal-ITP approval
1.0
Develop Customer Success Philosophy - - [x] Information gather current outreach procedures and initiatives with @o-ram - [ ] Document functional duties for outreach #290 - [x] #247 - [x] Research best approaches to customer success philosophy in praxis - [x] Draft customer success philosophy - [x] Draft review by @o-ram @AnthonyRollins @laneymangan Jess and any other relevant parties - [x] Edit and provide second draft - [x] Second draft review - [x] Submit draft for Cal-ITP approval
non_test
develop customer success philosophy information gather current outreach procedures and initiatives with o ram document functional duties for outreach research best approaches to customer success philosophy in praxis draft customer success philosophy draft review by o ram anthonyrollins laneymangan jess and any other relevant parties edit and provide second draft second draft review submit draft for cal itp approval
0
193,128
14,640,498,312
IssuesEvent
2020-12-25 02:18:25
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
influxdata/influxdb: chronograf/bolt/users_test.go; 3 LoC
fresh test tiny
Found a possible issue in [influxdata/influxdb](https://www.github.com/influxdata/influxdb) at [chronograf/bolt/users_test.go](https://github.com/influxdata/influxdb/blob/ab87c23be6c630754787dcd9113cd86bd6afaaf1/chronograf/bolt/users_test.go#L552-L554) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to u at line 553 may start a goroutine [Click here to see the code in its original context.](https://github.com/influxdata/influxdb/blob/ab87c23be6c630754787dcd9113cd86bd6afaaf1/chronograf/bolt/users_test.go#L552-L554) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, u := range tt.users { s.Add(tt.ctx, &u) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following paths through the callgraph could lead to a goroutine: (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (MeasurementExists, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (createTagSetIterators, 5) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (All, 1) -> (ListTasks, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) (Add, 2) -> (SetUserPerms, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) -> (compactCache, 0) -> (WriteSnapshot, 0) -> (writeSnapshotAndCommit, 3) -> (WriteSnapshot, 1) -> (NextGeneration, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (measurementNamesByTagFilter, 5) -> (valEqual, 1) -> (HasTagKey, 2) (Add, 2) -> (Query, 2) -> (Query, 3) -> (Any, 2) -> (executeQuery, 5) -> (ExecuteStatement, 3) -> (executeShowTagValues, 3) -> (getDefaultRP, 3) -> (DeleteSeriesRange, 3) -> (DeleteSeriesRangeWithPredicate, 2) -> (deleteSeriesRange, 3) -> (cleanupMeasurement, 1) -> (DeleteWithLock, 2) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) (Add, 2) -> (Create, 2) -> (CreateOrganization, 2) -> (Errorf, 2) -> (CreateDatabaseWithRetentionPolicy, 2) -> (CreateRetentionPolicy, 3) -> (ErrDatabaseNotFound, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) -> (compactCache, 0) -> (WriteSnapshot, 0) -> (writeSnapshotAndCommit, 3) -> (WriteSnapshot, 1) (Add, 2) -> (Generate, 0) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (FormatUint, 2) -> (detectFast, 0) -> (sortDedupStrings, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (NewService, 2) -> (location, 2) -> (Ping, 1) -> (pingTimeout, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (IteratorCost, 2) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (NewStringSliceIterator, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (All, 1) -> (Roles, 2) -> (roles, 2) (Add, 2) -> (Write, 2) -> (Write, 1) -> (PutUvarint, 2) -> (Open, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (LoadMetadataIndex, 2) -> (Uint64, 2) -> (saveNoLock, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) -> (PrecreateShardGroups, 2) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (NewService, 2) -> (location, 2) -> (Ping, 1) -> (pingTimeout, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (Commit, 0) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) (Add, 2) -> (Write, 2) -> (Encode, 1) -> (Float64bits, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Delete, 1) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Create, 2) -> (New, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (cursorHasData, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Flush, 1) -> (WritePoints, 4) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (MeasurementExists, 1) (Add, 2) -> (All, 1) -> (ListTasks, 1) -> (close, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Query, 2) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (FormatUint, 2) -> (VisitorFn, 1) -> (sortDedupStrings, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Update, 2) -> (New, 1) (Add, 2) -> (CreateUser, 3) -> (Clone, 0) -> (CloneUsers, 0) -> (Open, 1) (Add, 2) -> (Create, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (Open, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (cursorHasData, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (All, 1) -> (Users, 2) -> (encodeJSON, 4) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) (Add, 2) -> (Create, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (walkShards, 2) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (Flush, 1) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (LoadMetadataIndex, 2) -> (Uint64, 2) -> (saveNoLock, 0) -> (ForEachField, 1) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (NewParallelMergeIterator, 3) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (Create, 2) -> (CreateUser, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (AddRoleUsers, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Write, 2) -> (write, 2) -> (newEntryValues, 1) -> (purge, 0) (Add, 2) -> (All, 1) -> (ForEach, 1) -> (SeriesIDSet, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Flush, 1) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (WindowAggregate, 2) -> (NewWindowAggregateResultSet, 3) -> (newMultiShardArrayCursors, 4) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) (Add, 2) -> (Update, 2) -> (Delete, 2) -> (FindUser, 2) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Query, 2) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (copy, 2) -> (Base, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) -> (Enabled, 2) -> (add, 1) -> (purge, 0) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (Deduplicate, 0) -> (Stable, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (ReadFilter, 2) -> (NewFilteredResultSet, 4) -> (newMultiShardArrayCursors, 4) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (apply, 1) (Add, 2) -> (Query, 2) -> (Encode, 1) -> (Float64bits, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (put, 3) -> (Put, 4) -> (copy, 2) -> (Base, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (SetRolePerms, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Create, 2) -> (ID, 0) -> (MustIDBase16, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (CreateBucket, 2) -> (Diff, 3) -> (NewRetentionPolicyInfo, 0) -> (DefaultRetentionPolicyInfo, 0) (Add, 2) -> (Create, 2) -> (FindBucketByID, 2) -> (Errorf, 2) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) -> (replace, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableLevelCompactions, 1) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) -> (Enabled, 2) -> (Open, 1) (Add, 2) -> (Update, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (Open, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (ReadGroup, 2) -> (NewGroupResultSet, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (Write, 1) -> (PutUvarint, 2) -> (Open, 1) (Add, 2) -> (CreateUser, 3) -> (ID, 0) -> (MustIDBase16, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (CreateUserResourceMapping, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (add, 1) -> (purge, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (init, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (Open, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (createSeriesListIfNotExists, 2) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Write, 2) -> (writePoint, 2) -> (CreateDB, 2) (Add, 2) -> (Create, 2) -> (CreateBucket, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (Query, 2) -> (Query, 3) -> (Any, 2) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (Rollback, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableLevelCompactions, 1) -> (compact, 1) -> (Release, 1) (Add, 2) -> (Create, 2) -> (CreateOrganization, 2) -> (Errorf, 2) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (CreateRole, 2) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Update, 2) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (Quote, 1) -> (VisitorFn, 1) -> (sortDedupStrings, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) -> (replace, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) -> (PrecreateShardGroups, 2) -> (createSeriesListIfNotExists, 2) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (walkShards, 2) (Add, 2) -> (Write, 2) -> (nextDataNode, 0) -> (Next, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) (Add, 2) -> (Error, 0) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (Quote, 1) -> (VisitorFn, 1) -> (sortDedupStrings, 1) ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ab87c23be6c630754787dcd9113cd86bd6afaaf1
1.0
influxdata/influxdb: chronograf/bolt/users_test.go; 3 LoC - Found a possible issue in [influxdata/influxdb](https://www.github.com/influxdata/influxdb) at [chronograf/bolt/users_test.go](https://github.com/influxdata/influxdb/blob/ab87c23be6c630754787dcd9113cd86bd6afaaf1/chronograf/bolt/users_test.go#L552-L554) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to u at line 553 may start a goroutine [Click here to see the code in its original context.](https://github.com/influxdata/influxdb/blob/ab87c23be6c630754787dcd9113cd86bd6afaaf1/chronograf/bolt/users_test.go#L552-L554) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, u := range tt.users { s.Add(tt.ctx, &u) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following paths through the callgraph could lead to a goroutine: (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (MeasurementExists, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (createTagSetIterators, 5) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (All, 1) -> (ListTasks, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) (Add, 2) -> (SetUserPerms, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) -> (compactCache, 0) -> (WriteSnapshot, 0) -> (writeSnapshotAndCommit, 3) -> (WriteSnapshot, 1) -> (NextGeneration, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (measurementNamesByTagFilter, 5) -> (valEqual, 1) -> (HasTagKey, 2) (Add, 2) -> (Query, 2) -> (Query, 3) -> (Any, 2) -> (executeQuery, 5) -> (ExecuteStatement, 3) -> (executeShowTagValues, 3) -> (getDefaultRP, 3) -> (DeleteSeriesRange, 3) -> (DeleteSeriesRangeWithPredicate, 2) -> (deleteSeriesRange, 3) -> (cleanupMeasurement, 1) -> (DeleteWithLock, 2) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) (Add, 2) -> (Create, 2) -> (CreateOrganization, 2) -> (Errorf, 2) -> (CreateDatabaseWithRetentionPolicy, 2) -> (CreateRetentionPolicy, 3) -> (ErrDatabaseNotFound, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) -> (compactCache, 0) -> (WriteSnapshot, 0) -> (writeSnapshotAndCommit, 3) -> (WriteSnapshot, 1) (Add, 2) -> (Generate, 0) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (FormatUint, 2) -> (detectFast, 0) -> (sortDedupStrings, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (NewService, 2) -> (location, 2) -> (Ping, 1) -> (pingTimeout, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (IteratorCost, 2) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (NewStringSliceIterator, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (All, 1) -> (Roles, 2) -> (roles, 2) (Add, 2) -> (Write, 2) -> (Write, 1) -> (PutUvarint, 2) -> (Open, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (LoadMetadataIndex, 2) -> (Uint64, 2) -> (saveNoLock, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) -> (PrecreateShardGroups, 2) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (NewService, 2) -> (location, 2) -> (Ping, 1) -> (pingTimeout, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (Commit, 0) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) (Add, 2) -> (Write, 2) -> (Encode, 1) -> (Float64bits, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Delete, 1) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Create, 2) -> (New, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (cursorHasData, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Flush, 1) -> (WritePoints, 4) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (MeasurementExists, 1) (Add, 2) -> (All, 1) -> (ListTasks, 1) -> (close, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Query, 2) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (FormatUint, 2) -> (VisitorFn, 1) -> (sortDedupStrings, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Update, 2) -> (New, 1) (Add, 2) -> (CreateUser, 3) -> (Clone, 0) -> (CloneUsers, 0) -> (Open, 1) (Add, 2) -> (Create, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (Open, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (tagKeysWithFieldPredicate, 3) -> (cursorHasData, 1) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (All, 1) -> (Users, 2) -> (encodeJSON, 4) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableSnapshotCompactions, 0) (Add, 2) -> (Create, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (walkShards, 2) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (Flush, 1) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (LoadMetadataIndex, 2) -> (Uint64, 2) -> (saveNoLock, 0) -> (ForEachField, 1) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (NewParallelMergeIterator, 3) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (Create, 2) -> (CreateUser, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (AddRoleUsers, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Write, 2) -> (write, 2) -> (newEntryValues, 1) -> (purge, 0) (Add, 2) -> (All, 1) -> (ForEach, 1) -> (SeriesIDSet, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Flush, 1) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (WindowAggregate, 2) -> (NewWindowAggregateResultSet, 3) -> (newMultiShardArrayCursors, 4) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) (Add, 2) -> (Update, 2) -> (Delete, 2) -> (FindUser, 2) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Query, 2) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (copy, 2) -> (Base, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) -> (Enabled, 2) -> (add, 1) -> (purge, 0) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (Deduplicate, 0) -> (Stable, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (ReadFilter, 2) -> (NewFilteredResultSet, 4) -> (newMultiShardArrayCursors, 4) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (apply, 1) (Add, 2) -> (Query, 2) -> (Encode, 1) -> (Float64bits, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (put, 3) -> (Put, 4) -> (copy, 2) -> (Base, 1) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (newParallelIterator, 1) -> (newBooleanParallelIterator, 1) (Add, 2) -> (SetRolePerms, 3) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (writeToLog, 1) -> (scheduleSync, 0) (Add, 2) -> (Create, 2) -> (ID, 0) -> (MustIDBase16, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (CreateBucket, 2) -> (Diff, 3) -> (NewRetentionPolicyInfo, 0) -> (DefaultRetentionPolicyInfo, 0) (Add, 2) -> (Create, 2) -> (FindBucketByID, 2) -> (Errorf, 2) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Query, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) -> (replace, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableLevelCompactions, 1) (Add, 2) -> (Write, 2) -> (ServeHTTP, 2) -> (Enabled, 2) -> (Open, 1) (Add, 2) -> (Update, 2) -> (New, 1) -> (NewIDGenerator, 0) -> (Open, 1) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (ReadGroup, 2) -> (NewGroupResultSet, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (Write, 1) -> (PutUvarint, 2) -> (Open, 1) (Add, 2) -> (CreateUser, 3) -> (ID, 0) -> (MustIDBase16, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (add, 1) -> (purge, 0) (Add, 2) -> (Create, 2) -> (CreateUserResourceMapping, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (add, 1) -> (purge, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (init, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) (Add, 2) -> (Create, 2) -> (Run, 2) -> (Fatalf, 2) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagKeys, 2) -> (NewStringSliceIterator, 1) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (MeasurementNamesByPredicate, 1) -> (MergeSortedBytes, 1) -> (MergeSortedBytes, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Write, 2) -> (Values, 1) -> (Deduplicate, 0) -> (Open, 1) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (openV1Meta, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (createSeriesListIfNotExists, 2) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Write, 2) -> (writePoint, 2) -> (CreateDB, 2) (Add, 2) -> (Create, 2) -> (CreateBucket, 2) -> (Diff, 3) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Update, 2) -> (Do, 1) -> (Release, 0) -> (newIndexSeriesCursor, 3) -> (newIndexSeriesCursorInfluxQLPred, 3) -> (FieldKeysByPredicate, 1) -> (MeasurementNamesByPredicate, 1) -> (MeasurementNamesByPredicate, 2) -> (measurementNamesByPredicate, 2) -> (measurementNamesByTagPredicate, 5) -> (checkMeasurement, 2) -> (measurementHasEmptyTagValue, 3) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (TagValues, 2) -> (tagValues, 3) -> (TagValues, 3) -> (mergeTagValues, 2) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (Query, 2) -> (Query, 3) -> (Any, 2) (Add, 2) -> (All, 1) -> (Load, 1) -> (Delete, 1) -> (Rollback, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) -> (enableLevelCompactions, 1) -> (compact, 1) -> (Release, 1) (Add, 2) -> (Create, 2) -> (CreateOrganization, 2) -> (Errorf, 2) -> (WritePoints, 4) -> (NewPoint, 4) -> (pointKey, 4) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 0) -> (Load, 0) -> (writePoints, 1) -> (WritePoints, 3) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) -> (tagValuesByKeyAndExpr, 4) -> (seriesByExprIterator, 2) -> (hasTagKey, 2) -> (HasTagKey, 2) (Add, 2) -> (CreateRole, 2) -> (Post, 4) -> (Do, 6) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) (Add, 2) -> (Query, 2) -> (Do, 1) -> (Release, 0) -> (TagKeys, 3) -> (MeasurementTagKeyValuesByExpr, 5) (Add, 2) -> (Update, 2) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (Quote, 1) -> (VisitorFn, 1) -> (sortDedupStrings, 1) (Add, 2) -> (Create, 2) -> (Do, 1) -> (Release, 0) -> (measurementFields, 2) -> (MergeSortedStrings, 1) -> (MergeSortedStrings, 2) -> (MeasurementNamesByRegex, 1) -> (MeasurementNamesByRegex, 1) -> (fetchByteValues, 1) (Add, 2) -> (Put, 2) -> (Req, 3) -> (Headers, 1) -> (SetCompactionsEnabled, 1) -> (Engine, 0) -> (engineNoLock, 0) -> (free, 0) -> (writeSnapshotAndCommit, 3) -> (Replace, 2) -> (replace, 3) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (Open, 1) -> (runPrecreation, 1) -> (precreate, 1) -> (PrecreateShardGroups, 2) -> (createSeriesListIfNotExists, 2) -> (CheckLogFile, 0) -> (checkLogFile, 0) (Add, 2) -> (Put, 2) -> (Close, 0) -> (cancel, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) -> (walkShards, 2) (Add, 2) -> (Write, 2) -> (nextDataNode, 0) -> (Next, 0) -> (add, 1) -> (purge, 0) (Add, 2) -> (Put, 2) -> (, 0) -> (mapValuer, 1) (Add, 2) -> (Error, 0) -> (String, 0) -> (RUnlock, 0) -> (Visit, 1) -> (Quote, 1) -> (VisitorFn, 1) -> (sortDedupStrings, 1) ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ab87c23be6c630754787dcd9113cd86bd6afaaf1
test
influxdata influxdb chronograf bolt users test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to u at line may start a goroutine click here to show the line s of go which triggered the analyzer go for u range tt users s add tt ctx u click here to show extra information the analyzer produced the following paths through the callgraph could lead to a goroutine add put close cancel tagvalues tagvalues tagvalues mergetagvalues measurementexists add put close cancel tagvalues tagvalues tagvalues mergetagvalues createtagsetiterators add update do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex fetchbytevalues add create do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex newparalleliterator newbooleanparalleliterator add update do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex newparalleliterator newbooleanparalleliterator add all listtasks add all load delete add setuserperms post do add create do release add put close cancel enablesnapshotcompactions compactcache writesnapshot writesnapshotandcommit writesnapshot nextgeneration add put close cancel tagkeys newstringsliceiterator measurementtagkeyvaluesbyexpr tagvaluesbykeyandexpr measurementnamesbytagfilter valequal hastagkey add query query any executequery executestatement executeshowtagvalues getdefaultrp deleteseriesrange deleteseriesrangewithpredicate deleteseriesrange cleanupmeasurement deletewithlock add write servehttp add create createorganization errorf createdatabasewithretentionpolicy createretentionpolicy errdatabasenotfound add put close cancel enablesnapshotcompactions compactcache writesnapshot writesnapshotandcommit writesnapshot add generate string runlock visit formatuint detectfast sortdedupstrings add put mapvaluer newservice location ping pingtimeout add put close cancel tagvalues tagvalues tagvalues mergetagvalues iteratorcost add create do release tagkeyswithfieldpredicate newstringsliceiterator fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add all roles roles add write write putuvarint open add update do release tagkeys measurementtagkeyvaluesbyexpr add create run fatalf loadmetadataindex savenolock add put mapvaluer open runprecreation precreate precreateshardgroups add query mapvaluer newservice location ping pingtimeout add all load delete commit add query mapvaluer open add create run fatalf open load writepoints add create run fatalf open add write encode open add put mapvaluer delete writetolog schedulesync add create new add put mapvaluer open add put mapvaluer open runprecreation precreate add create do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex add update do release tagkeys measurementtagkeyvaluesbyexpr tagvaluesbykeyandexpr seriesbyexpriterator hastagkey hastagkey add update do release tagkeyswithfieldpredicate cursorhasdata fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add put mapvaluer flush writepoints add put mapvaluer open load add create do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex measurementexists add all listtasks close add purge add create run fatalf open load add query string runlock visit formatuint visitorfn sortdedupstrings add query mapvaluer open load writepoints writepoints checklogfile checklogfile add update new add createuser clone cloneusers open add create new newidgenerator open add query mapvaluer open load add query do release tagkeyswithfieldpredicate cursorhasdata fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add all users encodejson add put close cancel enablesnapshotcompactions add create new newidgenerator add purge add put req headers walkshards add put req headers setcompactionsenabled engine enginenolock free add put close cancel flush writepoints newpoint pointkey add create run fatalf loadmetadataindex savenolock foreachfield checklogfile checklogfile add put close cancel tagvalues tagvalues tagvalues mergetagvalues newparallelmergeiterator newparalleliterator newbooleanparalleliterator add create createuser diff writepoints newpoint pointkey add update do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex add query do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex fetchbytevalues add create do release newindexseriescursor newindexseriescursorinfluxqlpred fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add addroleusers post do add write write newentryvalues purge add all foreach seriesidset add purge add put mapvaluer flush writepoints newpoint pointkey add put close cancel windowaggregate newwindowaggregateresultset newmultishardarraycursors fieldkeysbypredicate measurementnamesbypredicate mergesortedbytes mergesortedbytes measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add query mapvaluer add put mapvaluer open runprecreation add update delete finduser writetolog schedulesync add query add put mapvaluer copy base add create do release tagkeys measurementtagkeyvaluesbyexpr add put req headers setcompactionsenabled engine add write servehttp enabled add purge add write values deduplicate deduplicate stable add put mapvaluer setcompactionsenabled engine enginenolock free writesnapshotandcommit replace add put close cancel readfilter newfilteredresultset newmultishardarraycursors fieldkeysbypredicate measurementnamesbypredicate mergesortedbytes mergesortedbytes measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add write values deduplicate apply add query encode open add put put put copy base add query mapvaluer open add put close cancel tagkeys newstringsliceiterator measurementtagkeyvaluesbyexpr add put mapvaluer open load writepoints add put mapvaluer open load writepoints add query do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex newparalleliterator newbooleanparalleliterator add setroleperms post do add put mapvaluer open add all load delete writetolog schedulesync add create id add purge add create createbucket diff newretentionpolicyinfo defaultretentionpolicyinfo add create findbucketbyid errorf writepoints newpoint pointkey add query mapvaluer open load writepoints add put mapvaluer setcompactionsenabled engine enginenolock free writesnapshotandcommit replace replace add put mapvaluer open add put mapvaluer open load add put mapvaluer add purge add query do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex add put close cancel enablelevelcompactions add write servehttp enabled open add update new newidgenerator open add put close cancel readgroup newgroupresultset newindexseriescursorinfluxqlpred fieldkeysbypredicate measurementnamesbypredicate mergesortedbytes mergesortedbytes measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add put write putuvarint open add createuser id add purge add create do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex fetchbytevalues add all load delete add purge add create createuserresourcemapping diff writepoints newpoint pointkey add create run fatalf add purge add query do release newindexseriescursor newindexseriescursorinfluxqlpred fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add write init add put mapvaluer setcompactionsenabled engine enginenolock free add create run fatalf open load writepoints writepoints checklogfile checklogfile add put close cancel tagkeys newstringsliceiterator measurementtagkeyvaluesbyexpr tagvaluesbykeyandexpr measurementnamesbypredicate mergesortedbytes mergesortedbytes measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add write values deduplicate open add put mapvaluer open load writepoints createserieslistifnotexists checklogfile checklogfile add write writepoint createdb add create createbucket diff writepoints newpoint pointkey add update do release newindexseriescursor newindexseriescursorinfluxqlpred fieldkeysbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbypredicate measurementnamesbytagpredicate checkmeasurement measurementhasemptytagvalue add put close cancel tagvalues tagvalues tagvalues mergetagvalues measurementnamesbyregex fetchbytevalues add create do release tagkeys measurementtagkeyvaluesbyexpr tagvaluesbykeyandexpr seriesbyexpriterator hastagkey hastagkey add query query any add all load delete rollback add put close cancel enablelevelcompactions compact release add create createorganization errorf writepoints newpoint pointkey add put req headers setcompactionsenabled engine enginenolock free writesnapshotandcommit replace add put mapvaluer open load writepoints writepoints checklogfile checklogfile add query do release tagkeys measurementtagkeyvaluesbyexpr tagvaluesbykeyandexpr seriesbyexpriterator hastagkey hastagkey add createrole post do add put mapvaluer setcompactionsenabled engine add query do release tagkeys measurementtagkeyvaluesbyexpr add update string runlock visit quote visitorfn sortdedupstrings add create do release measurementfields mergesortedstrings mergesortedstrings measurementnamesbyregex measurementnamesbyregex fetchbytevalues add put req headers setcompactionsenabled engine enginenolock free writesnapshotandcommit replace replace add put mapvaluer open runprecreation precreate precreateshardgroups createserieslistifnotexists checklogfile checklogfile add put close cancel add put mapvaluer walkshards add write nextdatanode next add purge add put mapvaluer add error string runlock visit quote visitorfn sortdedupstrings leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
264,893
23,145,072,309
IssuesEvent
2022-07-28 23:13:23
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Seridores - Registro por lotação - Capitão Enéas
generalization test development template-Síntese tecnologia informatica tag-Servidores subtag-Registro por lotação
DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Capitão Enéas.
1.0
Teste de generalizacao para a tag Seridores - Registro por lotação - Capitão Enéas - DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Capitão Enéas.
test
teste de generalizacao para a tag seridores registro por lotação capitão enéas dod realizar o teste de generalização do validador da tag seridores registro por lotação para o município de capitão enéas
1
77,013
9,976,937,797
IssuesEvent
2019-07-09 16:02:40
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Caching documentation contains instruction for removed Guava support
status: forward-port type: documentation
Forward port of issue #17446 to 2.2.0.M5.
1.0
Caching documentation contains instruction for removed Guava support - Forward port of issue #17446 to 2.2.0.M5.
non_test
caching documentation contains instruction for removed guava support forward port of issue to
0
175,621
21,313,871,898
IssuesEvent
2022-04-16 01:13:28
AlexRogalskiy/java-patterns
https://api.github.com/repos/AlexRogalskiy/java-patterns
opened
CVE-2020-7774 (High) detected in y18n-3.2.1.tgz
security vulnerability
## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-3.2.1.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz">https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/cacache/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npm-5.1.0.tgz - cacache-9.2.9.tgz - :x: **y18n-3.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7774 (High) detected in y18n-3.2.1.tgz - ## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-3.2.1.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz">https://registry.npmjs.org/y18n/-/y18n-3.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/cacache/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npm-5.1.0.tgz - cacache-9.2.9.tgz - :x: **y18n-3.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tgz cve high severity vulnerability vulnerable library tgz the bare bones internationalization library used by yargs library home page a href path to dependency file package json path to vulnerable library node modules npm node modules cacache node modules package json dependency hierarchy gitbook cli tgz root library npm tgz cacache tgz x tgz vulnerable library found in base branch master vulnerability details this affects the package before and poc by const require setlocale proto updatelocale polluted true console log polluted true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
324,216
9,886,301,046
IssuesEvent
2019-06-25 06:18:18
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.pixiv.net - design is broken
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.pixiv.net/member_illust.php?mode=medium **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: <PR> black bar on bottom of page **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/4b1fc6ae-fb3f-4974-bbfa-a0b006c7caa2-thumb.jpg)](https://webcompat.com/uploads/2019/6/4b1fc6ae-fb3f-4974-bbfa-a0b006c7caa2.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.pixiv.net - design is broken - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.pixiv.net/member_illust.php?mode=medium **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: <PR> black bar on bottom of page **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/4b1fc6ae-fb3f-4974-bbfa-a0b006c7caa2-thumb.jpg)](https://webcompat.com/uploads/2019/6/4b1fc6ae-fb3f-4974-bbfa-a0b006c7caa2.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
design is broken url browser version firefox mobile operating system android tested another browser yes problem type design is broken description black bar on bottom of page steps to reproduce browser configuration none from with ❤️
0
236,398
18,095,823,515
IssuesEvent
2021-09-22 08:54:13
arduino/arduino-cli
https://api.github.com/repos/arduino/arduino-cli
closed
Documentation: Unclear when a library is compiled as "new-style" or "legacy"
type: enhancement topic: documentation
## Bug Report The [library specification](https://arduino.github.io/arduino-cli/latest/library-specification/) documents two flavours of libraries: The modern one with source files in the `src/` directory, and the legacy one with source files in the root and `utility` directory. However, it is not immediately clear how the Arduino toolchain decides whether a library is one or the other. I thought that it depended on the presence of a `library.properties` file, which is also implied by the last section (but not 100% clear): https://arduino.github.io/arduino-cli/latest/library-specification/#old-library-format-pre-15 However, the section about [source code layout](https://arduino.github.io/arduino-cli/latest/library-specification/#source-code) is where this distinction is mainly made, but that has nothing explicit about when one or the other is chosen (but it seems to imply that the presence of a `src` directory might be the trigger? It would be good to clarify this behavior, maybe make it more explicit in the last section and refer to that from the "source" section? Maybe it would also be useful to clarify what "legacy" mode means exactly, i.e. make explicit that for all `library.properties` fields, whether the default value is used (i.e. like when the field is omitted in a new-style library), or whether there is some other default (e.g. like the name field which is probably deduced from the folder name).
1.0
Documentation: Unclear when a library is compiled as "new-style" or "legacy" - ## Bug Report The [library specification](https://arduino.github.io/arduino-cli/latest/library-specification/) documents two flavours of libraries: The modern one with source files in the `src/` directory, and the legacy one with source files in the root and `utility` directory. However, it is not immediately clear how the Arduino toolchain decides whether a library is one or the other. I thought that it depended on the presence of a `library.properties` file, which is also implied by the last section (but not 100% clear): https://arduino.github.io/arduino-cli/latest/library-specification/#old-library-format-pre-15 However, the section about [source code layout](https://arduino.github.io/arduino-cli/latest/library-specification/#source-code) is where this distinction is mainly made, but that has nothing explicit about when one or the other is chosen (but it seems to imply that the presence of a `src` directory might be the trigger? It would be good to clarify this behavior, maybe make it more explicit in the last section and refer to that from the "source" section? Maybe it would also be useful to clarify what "legacy" mode means exactly, i.e. make explicit that for all `library.properties` fields, whether the default value is used (i.e. like when the field is omitted in a new-style library), or whether there is some other default (e.g. like the name field which is probably deduced from the folder name).
non_test
documentation unclear when a library is compiled as new style or legacy bug report the documents two flavours of libraries the modern one with source files in the src directory and the legacy one with source files in the root and utility directory however it is not immediately clear how the arduino toolchain decides whether a library is one or the other i thought that it depended on the presence of a library properties file which is also implied by the last section but not clear however the section about is where this distinction is mainly made but that has nothing explicit about when one or the other is chosen but it seems to imply that the presence of a src directory might be the trigger it would be good to clarify this behavior maybe make it more explicit in the last section and refer to that from the source section maybe it would also be useful to clarify what legacy mode means exactly i e make explicit that for all library properties fields whether the default value is used i e like when the field is omitted in a new style library or whether there is some other default e g like the name field which is probably deduced from the folder name
0
232,647
17,790,078,653
IssuesEvent
2021-08-31 15:15:08
netotz/p-dispersion-problem
https://api.github.com/repos/netotz/p-dispersion-problem
opened
Update README
documentation enhancement
Update README with: * Screenshots and GIFs of the program running * Help for the command-line tool * General information about the problem
1.0
Update README - Update README with: * Screenshots and GIFs of the program running * Help for the command-line tool * General information about the problem
non_test
update readme update readme with screenshots and gifs of the program running help for the command line tool general information about the problem
0
31,693
26,008,383,598
IssuesEvent
2022-12-20 21:54:23
compdemocracy/polis
https://api.github.com/repos/compdemocracy/polis
closed
Build out ansible deploy scripts
⚒️ infrastructure
This has been a convo since way back in https://github.com/pol-is/polisServer/issues/222 As I understand it, the idea is to use docker and ansible together. @ballPointPenguin is working on stuff here: https://github.com/ballPointPenguin/polis-playbook (Pls do suggest better issue title or scoping)
1.0
Build out ansible deploy scripts - This has been a convo since way back in https://github.com/pol-is/polisServer/issues/222 As I understand it, the idea is to use docker and ansible together. @ballPointPenguin is working on stuff here: https://github.com/ballPointPenguin/polis-playbook (Pls do suggest better issue title or scoping)
non_test
build out ansible deploy scripts this has been a convo since way back in as i understand it the idea is to use docker and ansible together ballpointpenguin is working on stuff here pls do suggest better issue title or scoping
0
630,182
20,099,911,937
IssuesEvent
2022-02-07 01:50:49
jqyang42/ht_buses
https://api.github.com/repos/jqyang42/ht_buses
closed
Route unassignment should be 2-click
enhancement medium priority
Currently non-intuitive to have to restart assign mode to unassign
1.0
Route unassignment should be 2-click - Currently non-intuitive to have to restart assign mode to unassign
non_test
route unassignment should be click currently non intuitive to have to restart assign mode to unassign
0
17,694
5,477,295,629
IssuesEvent
2017-03-12 06:23:32
stan-dev/math
https://api.github.com/repos/stan-dev/math
opened
better derivative of lgamma and digamma stability
code cleanup feature
#### Summary: From @bgoodri on http://discourse.mc-stan.org/t/numerical-error-tanks-stepsize-other-possibilities/293/3 For the complete Gamma function, I was thinking the other day that Stan could do better. It is known https://www.vttoth.com/CMS/projects/41-the-lanczos-approximation1 how to get arbitrary precision for the log-Gamma function, which Boost basically already implements (with some additional edge cases). But in the Stan Math Library, we naively call the digamma function (in doubles) to compute the derivative, which can be less accurate. So, we could try differentiating the Lanczos / Godfrey approximation directly, which is not difficult, definitely faster, and possibly more accurate. #### Current Version: v2.14.0
1.0
better derivative of lgamma and digamma stability - #### Summary: From @bgoodri on http://discourse.mc-stan.org/t/numerical-error-tanks-stepsize-other-possibilities/293/3 For the complete Gamma function, I was thinking the other day that Stan could do better. It is known https://www.vttoth.com/CMS/projects/41-the-lanczos-approximation1 how to get arbitrary precision for the log-Gamma function, which Boost basically already implements (with some additional edge cases). But in the Stan Math Library, we naively call the digamma function (in doubles) to compute the derivative, which can be less accurate. So, we could try differentiating the Lanczos / Godfrey approximation directly, which is not difficult, definitely faster, and possibly more accurate. #### Current Version: v2.14.0
non_test
better derivative of lgamma and digamma stability summary from bgoodri on for the complete gamma function i was thinking the other day that stan could do better it is known how to get arbitrary precision for the log gamma function which boost basically already implements with some additional edge cases but in the stan math library we naively call the digamma function in doubles to compute the derivative which can be less accurate so we could try differentiating the lanczos godfrey approximation directly which is not difficult definitely faster and possibly more accurate current version
0
271,435
23,604,657,911
IssuesEvent
2022-08-24 07:10:20
Xmetalfanx/linuxSetup
https://api.github.com/repos/Xmetalfanx/linuxSetup
closed
Strawberry code refactoring
enhancement Needs Testing Priority
this is just a "remember to test" issue .... I know i need to test on Debian and Fedora
1.0
Strawberry code refactoring - this is just a "remember to test" issue .... I know i need to test on Debian and Fedora
test
strawberry code refactoring this is just a remember to test issue i know i need to test on debian and fedora
1
50,036
6,050,716,241
IssuesEvent
2017-06-12 21:42:51
openshift/origin-web-console
https://api.github.com/repos/openshift/origin-web-console
opened
Test flake: User adds an image stream to a project after creating a new project using the Import YAML tab should process and create the images in the image stream
kind/test-flake priority/P2
https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_web_console/1495/console ``` 1) User adds an image stream to a project after creating a new project using the Import YAML tab should process and create the images in the image stream - Expected false to be true. at catalogPage.processImageStream.then (/data/src/github.com/openshift/origin-web-console/test/integration/features/user_adds_imagestream_to_project.spec.js:38:79) at /data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/goog/base.js:1582:15 at webdriver.promise.ControlFlow.runInNewFrame_ (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:1654:20) at notify (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:465:12) at notifyAll (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:442:7) at resolve (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:420:7) at fulfill (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:535:5) at Object.webdriver.promise.asap (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:776:5) at newFrame.onComplete (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:1670:25) ```
1.0
Test flake: User adds an image stream to a project after creating a new project using the Import YAML tab should process and create the images in the image stream - https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_web_console/1495/console ``` 1) User adds an image stream to a project after creating a new project using the Import YAML tab should process and create the images in the image stream - Expected false to be true. at catalogPage.processImageStream.then (/data/src/github.com/openshift/origin-web-console/test/integration/features/user_adds_imagestream_to_project.spec.js:38:79) at /data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/goog/base.js:1582:15 at webdriver.promise.ControlFlow.runInNewFrame_ (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:1654:20) at notify (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:465:12) at notifyAll (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:442:7) at resolve (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:420:7) at fulfill (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:535:5) at Object.webdriver.promise.asap (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:776:5) at newFrame.onComplete (/data/src/github.com/openshift/origin-web-console/node_modules/selenium-webdriver/lib/webdriver/promise.js:1670:25) ```
test
test flake user adds an image stream to a project after creating a new project using the import yaml tab should process and create the images in the image stream user adds an image stream to a project after creating a new project using the import yaml tab should process and create the images in the image stream expected false to be true at catalogpage processimagestream then data src github com openshift origin web console test integration features user adds imagestream to project spec js at data src github com openshift origin web console node modules selenium webdriver lib goog base js at webdriver promise controlflow runinnewframe data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at notify data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at notifyall data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at resolve data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at fulfill data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at object webdriver promise asap data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js at newframe oncomplete data src github com openshift origin web console node modules selenium webdriver lib webdriver promise js
1
114,048
9,672,826,377
IssuesEvent
2019-05-22 05:27:38
SatelliteQE/automation-tools
https://api.github.com/repos/SatelliteQE/automation-tools
closed
Content-host test's failing due to timeout of 2mins
bug test-failure urgent
CH test failing due to firewall issues on the images and blanking out avahi-daemon responses ``` <lpramuk> sjagtap: avahi-daemon just starts up before firewalld starts so it manages to broadcast for very short time, and record is discarded after 120sec of silence <lpramuk> sjagtap: solution is to set fw <lpramuk> sjagtap: what happens if firewalld starts before avahi - answer is obvious :) ```
1.0
Content-host test's failing due to timeout of 2mins - CH test failing due to firewall issues on the images and blanking out avahi-daemon responses ``` <lpramuk> sjagtap: avahi-daemon just starts up before firewalld starts so it manages to broadcast for very short time, and record is discarded after 120sec of silence <lpramuk> sjagtap: solution is to set fw <lpramuk> sjagtap: what happens if firewalld starts before avahi - answer is obvious :) ```
test
content host test s failing due to timeout of ch test failing due to firewall issues on the images and blanking out avahi daemon responses sjagtap avahi daemon just starts up before firewalld starts so it manages to broadcast for very short time and record is discarded after of silence sjagtap solution is to set fw sjagtap what happens if firewalld starts before avahi answer is obvious
1
133,693
10,855,776,780
IssuesEvent
2019-11-13 19:08:48
ethereum/solidity
https://api.github.com/repos/ethereum/solidity
closed
[soltest/isoltest] different test results on macOS & Linux
bug :bug: testing :hammer:
`soltest` & `isoltest` currently show different test results on macOS and Linux. CI should execute tests in the context of macOS. I did some changes to `.circleci/config.yml` in #7464 to show the different test results in CircleCI, but probably the way I did these changes is not optimal.
1.0
[soltest/isoltest] different test results on macOS & Linux - `soltest` & `isoltest` currently show different test results on macOS and Linux. CI should execute tests in the context of macOS. I did some changes to `.circleci/config.yml` in #7464 to show the different test results in CircleCI, but probably the way I did these changes is not optimal.
test
different test results on macos linux soltest isoltest currently show different test results on macos and linux ci should execute tests in the context of macos i did some changes to circleci config yml in to show the different test results in circleci but probably the way i did these changes is not optimal
1
217,558
16,855,806,710
IssuesEvent
2021-06-21 06:28:01
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
Case failureraftstore::test_region_heartbeat::test_region_heartbeat_timestamp
component/test
Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/40248/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/40248/display/redirect</a>
1.0
Case failureraftstore::test_region_heartbeat::test_region_heartbeat_timestamp - Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/40248/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/40248/display/redirect</a>
test
case failureraftstore test region heartbeat test region heartbeat timestamp latest build a href
1
163,817
6,206,342,780
IssuesEvent
2017-07-06 18:12:56
idaholab/raven
https://api.github.com/repos/idaholab/raven
opened
HDF5 limited variable inclusion
improvement priority_critical
-------- Issue Description -------- ##### What did you expect to see happen? Choice over what variables are included in HDF5 storage ##### What did you see instead? HDF5 stores anything it can perceive as being an input/output This has significant memory implications that are critical to a special assignment currently underway. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
HDF5 limited variable inclusion - -------- Issue Description -------- ##### What did you expect to see happen? Choice over what variables are included in HDF5 storage ##### What did you see instead? HDF5 stores anything it can perceive as being an input/output This has significant memory implications that are critical to a special assignment currently underway. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
non_test
limited variable inclusion issue description what did you expect to see happen choice over what variables are included in storage what did you see instead stores anything it can perceive as being an input output this has significant memory implications that are critical to a special assignment currently underway for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
0
15,286
19,286,431,409
IssuesEvent
2021-12-11 02:51:45
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Failed Requests counter not working
question log-processing
Hi, I have recently started using GoAccess. Everything is working great, except for one thing. Failed Requests counter always shows 0, but in status codes, I can see that there are many failed requests by status code. ![Screenshot 2021-09-02 at 08 15 15](https://user-images.githubusercontent.com/9365780/131850729-ca2d333e-6c1d-4d12-93ee-f0509f222a20.png) ![Screenshot 2021-09-02 at 08 15 33](https://user-images.githubusercontent.com/9365780/131850721-31708c87-3163-43d6-b2a9-c9126c85bd56.png) I tried to find some config related to this but with no luck. The version used is 1.5.1. Please advice...
1.0
Failed Requests counter not working - Hi, I have recently started using GoAccess. Everything is working great, except for one thing. Failed Requests counter always shows 0, but in status codes, I can see that there are many failed requests by status code. ![Screenshot 2021-09-02 at 08 15 15](https://user-images.githubusercontent.com/9365780/131850729-ca2d333e-6c1d-4d12-93ee-f0509f222a20.png) ![Screenshot 2021-09-02 at 08 15 33](https://user-images.githubusercontent.com/9365780/131850721-31708c87-3163-43d6-b2a9-c9126c85bd56.png) I tried to find some config related to this but with no luck. The version used is 1.5.1. Please advice...
non_test
failed requests counter not working hi i have recently started using goaccess everything is working great except for one thing failed requests counter always shows but in status codes i can see that there are many failed requests by status code i tried to find some config related to this but with no luck the version used is please advice
0
718,378
24,715,335,711
IssuesEvent
2022-10-20 06:22:56
application-research/estuary
https://api.github.com/repos/application-research/estuary
closed
Estuary gateways are incorrectly handling gzip types
Bug Priority: high
**Describe the bug** When pulling down gzip data through `dweb.link`, it's response content type is set to `Content-Type: application/gzip`. This is correct. When pulling down the same cid through `https://api.estuary.tech/gw/ipfs/{cid}`, the response content type is set to `Content-Type: application/x-gzip`. This is incorrect and causes issues with clients like `docker`, see: https://github.com/moby/moby/pull/25122 https://github.com/moby/moby/issues/27642 **To Reproduce** Steps to reproduce the behavior: *Working* 1. curl -Lv 'https://bafkreibn6rsjptxv2l54irdxe7osex574ceda5wco3ijvrqkzghlzzq2eq.ipfs.dweb.link' \ -H 'Authorization: ESTc524b2ce-ea3c-4ded-a846-3ffb5c0d376cARY' --output test.gz Note `content-type: application/gzip` *Broken* 1. curl -Lv 'https://api.estuary.tech/gw/ipfs/bafkreibn6rsjptxv2l54irdxe7osex574ceda5wco3ijvrqkzghlzzq2eq' \ -H 'Authorization: ESTc524b2ce-ea3c-4ded-a846-3ffb5c0d376cARY' --output test2.gz Note `content-type: application/x-gzip` **Expected behavior** Content type should be handled correctly, as it can affect third part applications. Estuary's handling of content types should be the same as `dweb.link` and return `content-type: application/gzip` for *.gz files. **Actual behavior** `content-type: application/x-gzip` for *.gz files. **Additional context** Prevents [IPCR](https://github.com/application-research/ipcr) and potentially other third party applications from using optimally performant estuary gateways
1.0
Estuary gateways are incorrectly handling gzip types - **Describe the bug** When pulling down gzip data through `dweb.link`, it's response content type is set to `Content-Type: application/gzip`. This is correct. When pulling down the same cid through `https://api.estuary.tech/gw/ipfs/{cid}`, the response content type is set to `Content-Type: application/x-gzip`. This is incorrect and causes issues with clients like `docker`, see: https://github.com/moby/moby/pull/25122 https://github.com/moby/moby/issues/27642 **To Reproduce** Steps to reproduce the behavior: *Working* 1. curl -Lv 'https://bafkreibn6rsjptxv2l54irdxe7osex574ceda5wco3ijvrqkzghlzzq2eq.ipfs.dweb.link' \ -H 'Authorization: ESTc524b2ce-ea3c-4ded-a846-3ffb5c0d376cARY' --output test.gz Note `content-type: application/gzip` *Broken* 1. curl -Lv 'https://api.estuary.tech/gw/ipfs/bafkreibn6rsjptxv2l54irdxe7osex574ceda5wco3ijvrqkzghlzzq2eq' \ -H 'Authorization: ESTc524b2ce-ea3c-4ded-a846-3ffb5c0d376cARY' --output test2.gz Note `content-type: application/x-gzip` **Expected behavior** Content type should be handled correctly, as it can affect third part applications. Estuary's handling of content types should be the same as `dweb.link` and return `content-type: application/gzip` for *.gz files. **Actual behavior** `content-type: application/x-gzip` for *.gz files. **Additional context** Prevents [IPCR](https://github.com/application-research/ipcr) and potentially other third party applications from using optimally performant estuary gateways
non_test
estuary gateways are incorrectly handling gzip types describe the bug when pulling down gzip data through dweb link it s response content type is set to content type application gzip this is correct when pulling down the same cid through the response content type is set to content type application x gzip this is incorrect and causes issues with clients like docker see to reproduce steps to reproduce the behavior working curl lv h authorization output test gz note content type application gzip broken curl lv h authorization output gz note content type application x gzip expected behavior content type should be handled correctly as it can affect third part applications estuary s handling of content types should be the same as dweb link and return content type application gzip for gz files actual behavior content type application x gzip for gz files additional context prevents and potentially other third party applications from using optimally performant estuary gateways
0
152,150
12,093,188,041
IssuesEvent
2020-04-19 18:36:00
gisellemartel/CONPASS
https://api.github.com/repos/gisellemartel/CONPASS
opened
AT - 21: (US4E - As a user, I want to be able to select/specify start and end rooms to and from any floor.)
Acceptance Test (SPRINT 4)
1. Launch the application 2. Click on the a building that has interior mode (Hall Building) and enter inter mode by tapping on the building 3. Once in interior mode, Click the direction button (blue button bottom right hand side) 4. Enter a starting point (on the floor you are currently on) and a destination Shortest path should be drawn
1.0
AT - 21: (US4E - As a user, I want to be able to select/specify start and end rooms to and from any floor.) - 1. Launch the application 2. Click on the a building that has interior mode (Hall Building) and enter inter mode by tapping on the building 3. Once in interior mode, Click the direction button (blue button bottom right hand side) 4. Enter a starting point (on the floor you are currently on) and a destination Shortest path should be drawn
test
at as a user i want to be able to select specify start and end rooms to and from any floor launch the application click on the a building that has interior mode hall building and enter inter mode by tapping on the building once in interior mode click the direction button blue button bottom right hand side enter a starting point on the floor you are currently on and a destination shortest path should be drawn
1
105,625
13,202,198,069
IssuesEvent
2020-08-14 11:47:13
IFRCGo/go-frontend
https://api.github.com/repos/IFRCGo/go-frontend
opened
[DESIGN] Validate the 3W Create and Edit button flows from emergency pages
UX - Design
Currently on emergency pages there is a 3W Create button, but the users get confused when they want to edit it. One issue is that the 3W tab has not been added yet to the emergency pages. But also, the flow of creating and editing 3Ws need to be clarified, since one user can create multiple 3Ws and want to edit them. The second issue is to make sure the 3W table is visible properly even on smaller screens, given the 3 buttons to edit get cut off when the table is too small and becomes scrollable. Also a dev issue to make sure the scroll bar is always visible.
1.0
[DESIGN] Validate the 3W Create and Edit button flows from emergency pages - Currently on emergency pages there is a 3W Create button, but the users get confused when they want to edit it. One issue is that the 3W tab has not been added yet to the emergency pages. But also, the flow of creating and editing 3Ws need to be clarified, since one user can create multiple 3Ws and want to edit them. The second issue is to make sure the 3W table is visible properly even on smaller screens, given the 3 buttons to edit get cut off when the table is too small and becomes scrollable. Also a dev issue to make sure the scroll bar is always visible.
non_test
validate the create and edit button flows from emergency pages currently on emergency pages there is a create button but the users get confused when they want to edit it one issue is that the tab has not been added yet to the emergency pages but also the flow of creating and editing need to be clarified since one user can create multiple and want to edit them the second issue is to make sure the table is visible properly even on smaller screens given the buttons to edit get cut off when the table is too small and becomes scrollable also a dev issue to make sure the scroll bar is always visible
0
218,201
16,964,764,402
IssuesEvent
2021-06-29 09:31:47
thesofproject/sof
https://api.github.com/repos/thesofproject/sof
opened
[BUG] JSL_RVP_NOCODEC check alsabat failed
Intel Linux Daily tests JSL bug
**Describe the bug** happened since inner daily test 4887 , while 4882 is still good when check-alsabat-playback ,get warning and the case failed ``` 2021-06-26 03:49:37 UTC [REMOTE_INFO] check the PCMs before alsabat test 2021-06-26 03:49:39 UTC [REMOTE_COMMAND] alsabat -Phw:sofnocodec,0 --standalone -n 240000 -r 48000 -c 2 -F 997 2021-06-26 03:49:40 UTC [REMOTE_COMMAND] alsabat -Chw:sofnocodec,0 -c 2 -r 48000 -F 997 WARNING: Signal too weak! WARNING: Found low peak 0.73 Hz, very close to DC FAIL: Peak freq too low 21.24 Hz FAIL: Peak freq too low 133.30 Hz FAIL: Peak freq too low 155.27 Hz FAIL: Peak freq too low 177.25 Hz FAIL: Peak freq too low 210.94 Hz FAIL: Peak freq too low 254.88 Hz FAIL: Peak freq too low 276.86 Hz FAIL: Peak freq too low 298.83 Hz FAIL: Peak freq too low 421.14 Hz alsa-utils version 1.2.2 Entering capture thread (ALSA). Get period size: 2048 buffer size: 16384 Recording ... Capture completed. BAT analysis: signal has 65536 frames at 48000 Hz, 2 channels, 2 bytes per sample. Channel 1 - Checking for target frequency 997.00 Hz Amplitude: 79.9; Percentage: [0] Detected peak at 0.73 Hz of 19.55 dB Total 27.4 dB from 0.73 to 15.38 Hz Detected peak at 21.24 Hz of 9.46 dB Total 27.4 dB from 21.24 to 21.24 Hz Detected peak at 133.30 Hz of 11.90 dB Total 27.6 dB from 132.57 to 133.30 Hz Detected peak at 155.27 Hz of 9.49 dB Total 27.7 dB from 155.27 to 155.27 Hz Detected peak at 177.25 Hz of 8.95 dB Total 27.8 dB from 177.25 to 177.25 Hz Detected peak at 210.94 Hz of 8.86 dB Total 27.8 dB from 210.94 to 210.94 Hz Detected peak at 254.88 Hz of 10.29 dB Total 27.9 dB from 254.88 to 254.88 Hz Detected peak at 276.86 Hz of 9.12 dB Total 27.9 dB from 276.86 to 276.86 Hz Detected peak at 298.83 Hz of 10.10 dB Total 28.0 dB from 298.83 to 298.83 Hz Detected peak at 421.14 Hz of 11.00 dB Total 28.1 dB from 421.14 to 421.14 Hz Detected at least 10 signal(s) in total Return value is -1003 ``` This is a regression , I use the sof recipe of inner daily test 4882 and it got a PASS back **To Reproduce** TPLG=sof-jsl-nocodec.tplg ~/sof-test/test-case/check-alsabat.sh -p hw:sofnocodec,0 -c hw:sofnocodec,0 -C 2 **Reproduction Rate** 100% **Environment** Kernel Branch: topic/sof-dev Kernel Commit: 5b851f48 SOF Branch: main SOF Commit: 2d70f100c7ad Topology: sof-jsl-nocodec.tplg Platform: jsl_rvp_nocodec
1.0
[BUG] JSL_RVP_NOCODEC check alsabat failed - **Describe the bug** happened since inner daily test 4887 , while 4882 is still good when check-alsabat-playback ,get warning and the case failed ``` 2021-06-26 03:49:37 UTC [REMOTE_INFO] check the PCMs before alsabat test 2021-06-26 03:49:39 UTC [REMOTE_COMMAND] alsabat -Phw:sofnocodec,0 --standalone -n 240000 -r 48000 -c 2 -F 997 2021-06-26 03:49:40 UTC [REMOTE_COMMAND] alsabat -Chw:sofnocodec,0 -c 2 -r 48000 -F 997 WARNING: Signal too weak! WARNING: Found low peak 0.73 Hz, very close to DC FAIL: Peak freq too low 21.24 Hz FAIL: Peak freq too low 133.30 Hz FAIL: Peak freq too low 155.27 Hz FAIL: Peak freq too low 177.25 Hz FAIL: Peak freq too low 210.94 Hz FAIL: Peak freq too low 254.88 Hz FAIL: Peak freq too low 276.86 Hz FAIL: Peak freq too low 298.83 Hz FAIL: Peak freq too low 421.14 Hz alsa-utils version 1.2.2 Entering capture thread (ALSA). Get period size: 2048 buffer size: 16384 Recording ... Capture completed. BAT analysis: signal has 65536 frames at 48000 Hz, 2 channels, 2 bytes per sample. Channel 1 - Checking for target frequency 997.00 Hz Amplitude: 79.9; Percentage: [0] Detected peak at 0.73 Hz of 19.55 dB Total 27.4 dB from 0.73 to 15.38 Hz Detected peak at 21.24 Hz of 9.46 dB Total 27.4 dB from 21.24 to 21.24 Hz Detected peak at 133.30 Hz of 11.90 dB Total 27.6 dB from 132.57 to 133.30 Hz Detected peak at 155.27 Hz of 9.49 dB Total 27.7 dB from 155.27 to 155.27 Hz Detected peak at 177.25 Hz of 8.95 dB Total 27.8 dB from 177.25 to 177.25 Hz Detected peak at 210.94 Hz of 8.86 dB Total 27.8 dB from 210.94 to 210.94 Hz Detected peak at 254.88 Hz of 10.29 dB Total 27.9 dB from 254.88 to 254.88 Hz Detected peak at 276.86 Hz of 9.12 dB Total 27.9 dB from 276.86 to 276.86 Hz Detected peak at 298.83 Hz of 10.10 dB Total 28.0 dB from 298.83 to 298.83 Hz Detected peak at 421.14 Hz of 11.00 dB Total 28.1 dB from 421.14 to 421.14 Hz Detected at least 10 signal(s) in total Return value is -1003 ``` This is a regression , I use the sof recipe of inner daily test 4882 and it got a PASS back **To Reproduce** TPLG=sof-jsl-nocodec.tplg ~/sof-test/test-case/check-alsabat.sh -p hw:sofnocodec,0 -c hw:sofnocodec,0 -C 2 **Reproduction Rate** 100% **Environment** Kernel Branch: topic/sof-dev Kernel Commit: 5b851f48 SOF Branch: main SOF Commit: 2d70f100c7ad Topology: sof-jsl-nocodec.tplg Platform: jsl_rvp_nocodec
test
jsl rvp nocodec check alsabat failed describe the bug happened since inner daily test while is still good when check alsabat playback get warning and the case failed utc check the pcms before alsabat test utc alsabat phw sofnocodec standalone n r c f utc alsabat chw sofnocodec c r f warning signal too weak warning found low peak hz very close to dc fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz fail peak freq too low hz alsa utils version entering capture thread alsa get period size buffer size recording capture completed bat analysis signal has frames at hz channels bytes per sample channel checking for target frequency hz amplitude percentage detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected peak at hz of db total db from to hz detected at least signal s in total return value is this is a regression i use the sof recipe of inner daily test and it got a pass back to reproduce tplg sof jsl nocodec tplg sof test test case check alsabat sh p hw sofnocodec c hw sofnocodec c reproduction rate environment kernel branch topic sof dev kernel commit sof branch main sof commit topology sof jsl nocodec tplg platform jsl rvp nocodec
1
343,415
30,664,140,550
IssuesEvent
2023-07-25 16:58:16
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
[BUG] Resource type management.cattle.io.clusterroletemplatebinding not found when trying to add cluster members to downstream clusters
kind/bug [zube]: To Test priority/0 status/release-blocker
**Rancher Server Setup** - Rancher version: `v2.7-head Commit ID: 95f0b50` - Installation option (Docker install/Helm Chart): `Helm` - If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc): `RKE1` - Proxy/Cert Details: `byo-valid` **Information about the Cluster** - Kubernetes version: `v1.26.6` - Cluster Type (Local/Downstream): `Downstream EC2 - 1 all-role node, 3 workers` - If downstream, what type of cluster? (Custom/Imported or specify provider for Hosted/Infrastructure Provider): `EC2` **User Information** - What is the role of the user logged in? (Admin/Cluster Owner/Cluster Member/Project Owner/Project Member/Custom): `Admin` **Describe the bug** When attempting to add a user as a cluster member for a downstream cluster, clicking the "Add" button leads to the "fail-whale" page with the following error message: `Resource type management.cattle.io.clusterroletemplatebinding not found` **To Reproduce** 1. Provision an EC2 HA cluster 2. Sign into the cluster 3. Create a downstream RKE1 EC2 cluster - 1 all-role, 3 workers 4. Once the cluster is up, create a new standard user - u1 5. Go into the downstream cluster -> Cluster and Project Members 6. Click the "Add" button for adding a cluster member **Result** Clicking the "Add" button for cluster members fails due to missing clusterroletemplatebinding **Expected Result** User is able to successfully add a cluster member with no errors **Screenshots** ![image](https://github.com/rancher/rancher/assets/107147272/5b56318d-df10-45e3-ad27-02529b065905) ![image](https://github.com/rancher/rancher/assets/107147272/f9099a11-5090-4bad-bdc1-3fd45006d61f) **Additional context** I can repro on both RKE1 and RKE2 clusters
1.0
[BUG] Resource type management.cattle.io.clusterroletemplatebinding not found when trying to add cluster members to downstream clusters - **Rancher Server Setup** - Rancher version: `v2.7-head Commit ID: 95f0b50` - Installation option (Docker install/Helm Chart): `Helm` - If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc): `RKE1` - Proxy/Cert Details: `byo-valid` **Information about the Cluster** - Kubernetes version: `v1.26.6` - Cluster Type (Local/Downstream): `Downstream EC2 - 1 all-role node, 3 workers` - If downstream, what type of cluster? (Custom/Imported or specify provider for Hosted/Infrastructure Provider): `EC2` **User Information** - What is the role of the user logged in? (Admin/Cluster Owner/Cluster Member/Project Owner/Project Member/Custom): `Admin` **Describe the bug** When attempting to add a user as a cluster member for a downstream cluster, clicking the "Add" button leads to the "fail-whale" page with the following error message: `Resource type management.cattle.io.clusterroletemplatebinding not found` **To Reproduce** 1. Provision an EC2 HA cluster 2. Sign into the cluster 3. Create a downstream RKE1 EC2 cluster - 1 all-role, 3 workers 4. Once the cluster is up, create a new standard user - u1 5. Go into the downstream cluster -> Cluster and Project Members 6. Click the "Add" button for adding a cluster member **Result** Clicking the "Add" button for cluster members fails due to missing clusterroletemplatebinding **Expected Result** User is able to successfully add a cluster member with no errors **Screenshots** ![image](https://github.com/rancher/rancher/assets/107147272/5b56318d-df10-45e3-ad27-02529b065905) ![image](https://github.com/rancher/rancher/assets/107147272/f9099a11-5090-4bad-bdc1-3fd45006d61f) **Additional context** I can repro on both RKE1 and RKE2 clusters
test
resource type management cattle io clusterroletemplatebinding not found when trying to add cluster members to downstream clusters rancher server setup rancher version head commit id installation option docker install helm chart helm if helm chart kubernetes cluster and version eks etc proxy cert details byo valid information about the cluster kubernetes version cluster type local downstream downstream all role node workers if downstream what type of cluster custom imported or specify provider for hosted infrastructure provider user information what is the role of the user logged in admin cluster owner cluster member project owner project member custom admin describe the bug when attempting to add a user as a cluster member for a downstream cluster clicking the add button leads to the fail whale page with the following error message resource type management cattle io clusterroletemplatebinding not found to reproduce provision an ha cluster sign into the cluster create a downstream cluster all role workers once the cluster is up create a new standard user go into the downstream cluster cluster and project members click the add button for adding a cluster member result clicking the add button for cluster members fails due to missing clusterroletemplatebinding expected result user is able to successfully add a cluster member with no errors screenshots additional context i can repro on both and clusters
1
344,988
30,779,336,316
IssuesEvent
2023-07-31 08:56:21
nerdvana-ro/infoarena
https://api.github.com/repos/nerdvana-ro/infoarena
opened
Schimbă cheia primară în ia_textblock_revision.
bug backend test
Cheia după `(name, timestamp)` duce la probleme în timpul testelor funcționale, cînd putem edita o pagină de două ori în aceeași secundă.
1.0
Schimbă cheia primară în ia_textblock_revision. - Cheia după `(name, timestamp)` duce la probleme în timpul testelor funcționale, cînd putem edita o pagină de două ori în aceeași secundă.
test
schimbă cheia primară în ia textblock revision cheia după name timestamp duce la probleme în timpul testelor funcționale cînd putem edita o pagină de două ori în aceeași secundă
1
285,849
24,702,276,360
IssuesEvent
2022-10-19 16:07:22
Kimthien94/Feedback-Online
https://api.github.com/repos/Kimthien94/Feedback-Online
opened
[BugID_71]_FUNC_Xóa topic gán cho trainer_Sau khi xóa thành công, Combobox [Lớp] hiển thị item mặc định
bug Fun_Incomplete Function comestic low Intergration test open
Precondition:Admin đứng tại màn hình Quản lý gán topic Chọn 1 item từ combobox [Lớp] Click vào button [Xóa] của 1 record nào đó Click vào button [Có] Thực tế: Combobox [Lớp] vẫn hiển thị item mặc định Mong đợi: Combobox [Lớp] vẫn hiển thị item đã chọn và refresh grid Testcase ID = 18
1.0
[BugID_71]_FUNC_Xóa topic gán cho trainer_Sau khi xóa thành công, Combobox [Lớp] hiển thị item mặc định - Precondition:Admin đứng tại màn hình Quản lý gán topic Chọn 1 item từ combobox [Lớp] Click vào button [Xóa] của 1 record nào đó Click vào button [Có] Thực tế: Combobox [Lớp] vẫn hiển thị item mặc định Mong đợi: Combobox [Lớp] vẫn hiển thị item đã chọn và refresh grid Testcase ID = 18
test
func xóa topic gán cho trainer sau khi xóa thành công combobox hiển thị item mặc định precondition admin đứng tại màn hình quản lý gán topic chọn item từ combobox click vào button của record nào đó click vào button thực tế combobox vẫn hiển thị item mặc định mong đợi combobox vẫn hiển thị item đã chọn và refresh grid testcase id
1
298,374
25,820,879,349
IssuesEvent
2022-12-12 09:29:13
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.4.0 - Alpha 1 - GCP Module testing
module/gcp module/cloud monitoring type/test/manual team/framework release test/4.4.0
# GCP module testing ## Modules tests information ||| |----------------------------------|------ | | **Main release candidate issue** | #15525 | | **Version** | 4.4.0 | | **Release candidate #** | Alpha 1 | | **Tag** | [v4.4.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.4.0-alpha1) | | **Previous modules tests issue** | - | The GCP module testing includes both `gcp-bucket` and `gcp-pubsub` integrations. ## Test report procedure All test results must have one the following statuses: | | | |-----------------|-------------------------------------------------------------------------| | :green_circle: | The test passed successfully. | | :yellow_circle: | The test passed but some enhancements need to be applied to the module. | | :red_circle: | The test failed. | | :white_circle: | The test does not apply to the given bucket type or service. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the Release 4.4.0 project for triage. The data available in our Wazuh-dev environment for each of these supported services will be used. Additional logs may be created if they are necessary or there is insufficient data for testing. ## Test results | Feature | Status | Issue ref. | |-------------------|---------|----------------| | Test using ossec.conf in a manager with Python 3.9 | :yellow_circle: | #15569 | | Test using ossec.conf in an agent with Python 3.6 | :green_circle: | - | ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @davidjiglesias - [x] @CarlosRS9 - [x] @fdalmaup
2.0
Release 4.4.0 - Alpha 1 - GCP Module testing - # GCP module testing ## Modules tests information ||| |----------------------------------|------ | | **Main release candidate issue** | #15525 | | **Version** | 4.4.0 | | **Release candidate #** | Alpha 1 | | **Tag** | [v4.4.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.4.0-alpha1) | | **Previous modules tests issue** | - | The GCP module testing includes both `gcp-bucket` and `gcp-pubsub` integrations. ## Test report procedure All test results must have one the following statuses: | | | |-----------------|-------------------------------------------------------------------------| | :green_circle: | The test passed successfully. | | :yellow_circle: | The test passed but some enhancements need to be applied to the module. | | :red_circle: | The test failed. | | :white_circle: | The test does not apply to the given bucket type or service. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the Release 4.4.0 project for triage. The data available in our Wazuh-dev environment for each of these supported services will be used. Additional logs may be created if they are necessary or there is insufficient data for testing. ## Test results | Feature | Status | Issue ref. | |-------------------|---------|----------------| | Test using ossec.conf in a manager with Python 3.9 | :yellow_circle: | #15569 | | Test using ossec.conf in an agent with Python 3.6 | :green_circle: | - | ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @davidjiglesias - [x] @CarlosRS9 - [x] @fdalmaup
test
release alpha gcp module testing gcp module testing modules tests information main release candidate issue version release candidate alpha tag previous modules tests issue the gcp module testing includes both gcp bucket and gcp pubsub integrations test report procedure all test results must have one the following statuses green circle the test passed successfully yellow circle the test passed but some enhancements need to be applied to the module red circle the test failed white circle the test does not apply to the given bucket type or service any failing test must be properly addressed with a new issue detailing the error and the possible cause it must be included in the release project for triage the data available in our wazuh dev environment for each of these supported services will be used additional logs may be created if they are necessary or there is insufficient data for testing test results feature status issue ref test using ossec conf in a manager with python yellow circle test using ossec conf in an agent with python green circle auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue davidjiglesias fdalmaup
1
106,928
9,197,546,888
IssuesEvent
2019-03-07 10:15:18
GateNLP/gateplugin-LearningFramework
https://api.github.com/repos/GateNLP/gateplugin-LearningFramework
closed
Xingyi's bug report
needs testing
I notice several bugs may need to remove in next version 1. Copy model from GPU machine to CPU only cause cuda error: raise RuntimeError(‘Attempting to deserialize object on a CUDA ’ (detail in slack) -- current solution is del state['optimizer'] in the __getstate__ seems removed the cuda error 2. Performance difference when copy model from GPU machine to CPU only machine. copied file: FileJsonPyTorch.model.module.pytorch about 10% accuracy difference in same test corpus. GATE version is same but pytorch is different 0.4.1 (gs8) vs 1.0.1 (xingyi's mac) 3. Path error when copy FileJsonPyTorch.model.wrapper.pickle - the meta file path is stored in FileJsonPyTorch.model.wrapper.pickle, this cause can't find file error when copy to different machine
1.0
Xingyi's bug report - I notice several bugs may need to remove in next version 1. Copy model from GPU machine to CPU only cause cuda error: raise RuntimeError(‘Attempting to deserialize object on a CUDA ’ (detail in slack) -- current solution is del state['optimizer'] in the __getstate__ seems removed the cuda error 2. Performance difference when copy model from GPU machine to CPU only machine. copied file: FileJsonPyTorch.model.module.pytorch about 10% accuracy difference in same test corpus. GATE version is same but pytorch is different 0.4.1 (gs8) vs 1.0.1 (xingyi's mac) 3. Path error when copy FileJsonPyTorch.model.wrapper.pickle - the meta file path is stored in FileJsonPyTorch.model.wrapper.pickle, this cause can't find file error when copy to different machine
test
xingyi s bug report i notice several bugs may need to remove in next version copy model from gpu machine to cpu only cause cuda error raise runtimeerror ‘attempting to deserialize object on a cuda ’ detail in slack current solution is del state in the getstate seems removed the cuda error performance difference when copy model from gpu machine to cpu only machine copied file filejsonpytorch model module pytorch about accuracy difference in same test corpus gate version is same but pytorch is different vs xingyi s mac path error when copy filejsonpytorch model wrapper pickle the meta file path is stored in filejsonpytorch model wrapper pickle this cause can t find file error when copy to different machine
1
7,430
3,973,572,599
IssuesEvent
2016-05-04 19:03:38
GitHubRGI/swagd
https://api.github.com/repos/GitHubRGI/swagd
opened
Ensure git lfs files work on Jenkins / Windows server
build test
See comment thread on Jenkins here: https://issues.jenkins-ci.org/browse/JENKINS-30318
1.0
Ensure git lfs files work on Jenkins / Windows server - See comment thread on Jenkins here: https://issues.jenkins-ci.org/browse/JENKINS-30318
non_test
ensure git lfs files work on jenkins windows server see comment thread on jenkins here
0
293,595
25,307,561,397
IssuesEvent
2022-11-17 15:11:12
vaadin/flow-components
https://api.github.com/repos/vaadin/flow-components
closed
Investigate failing LitRenderer ITs
test internal
### Description The following ITs started to fail: ``` 15:43:18  shouldRenderWithNoBoundValueProviders[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[]0> but was:<[Item: ]0> 15:43:18  org.junit.ComparisonFailure: expected:<[]0> but was:<[Item: ]0> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldRenderWithNoBoundValueProviders(LitRendererIT.java:48) ``` ``` 15:43:18  shouldRemoveTheRenderer[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[[object Object]]> but was:<[Item: 0]> 15:43:18  org.junit.ComparisonFailure: expected:<[[object Object]]> but was:<[Item: 0]> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldRemoveTheRenderer(LitRendererIT.java:55) ``` ``` 15:43:18  shouldSupportRendererInstanceSpecificProperties[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[Details: 0 (details)]> but was:<[]> 15:43:18  org.junit.ComparisonFailure: expected:<[Details: 0 (details)]> but was:<[]> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldSupportRendererInstanceSpecificProperties(LitRendererIT.java:119) ``` ### Expected outcome Tests should pass. ### Environment Vaadin version(s): 24.0 (latest master) ### Browsers Issue is not browser related
1.0
Investigate failing LitRenderer ITs - ### Description The following ITs started to fail: ``` 15:43:18  shouldRenderWithNoBoundValueProviders[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[]0> but was:<[Item: ]0> 15:43:18  org.junit.ComparisonFailure: expected:<[]0> but was:<[Item: ]0> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldRenderWithNoBoundValueProviders(LitRendererIT.java:48) ``` ``` 15:43:18  shouldRemoveTheRenderer[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[[object Object]]> but was:<[Item: 0]> 15:43:18  org.junit.ComparisonFailure: expected:<[[object Object]]> but was:<[Item: 0]> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldRemoveTheRenderer(LitRendererIT.java:55) ``` ``` 15:43:18  shouldSupportRendererInstanceSpecificProperties[ANY_Chrome_] 15:43:18  org.junit.ComparisonFailure: expected:<[Details: 0 (details)]> but was:<[]> 15:43:18  org.junit.ComparisonFailure: expected:<[Details: 0 (details)]> but was:<[]> at org.junit.Assert.assertEquals(Assert.java:117) at org.junit.Assert.assertEquals(Assert.java:146) at com.vaadin.flow.data.renderer.tests.LitRendererIT.shouldSupportRendererInstanceSpecificProperties(LitRendererIT.java:119) ``` ### Expected outcome Tests should pass. ### Environment Vaadin version(s): 24.0 (latest master) ### Browsers Issue is not browser related
test
investigate failing litrenderer its description the following its started to fail   shouldrenderwithnoboundvalueproviders   org junit comparisonfailure expected but was   org junit comparisonfailure expected but was at org junit assert assertequals assert java at org junit assert assertequals assert java at com vaadin flow data renderer tests litrendererit shouldrenderwithnoboundvalueproviders litrendererit java   shouldremovetherenderer   org junit comparisonfailure expected but was   org junit comparisonfailure expected but was at org junit assert assertequals assert java at org junit assert assertequals assert java at com vaadin flow data renderer tests litrendererit shouldremovetherenderer litrendererit java   shouldsupportrendererinstancespecificproperties   org junit comparisonfailure expected but was   org junit comparisonfailure expected but was at org junit assert assertequals assert java at org junit assert assertequals assert java at com vaadin flow data renderer tests litrendererit shouldsupportrendererinstancespecificproperties litrendererit java expected outcome tests should pass environment vaadin version s latest master browsers issue is not browser related
1
219,660
17,104,607,025
IssuesEvent
2021-07-09 15:47:27
moonfox/gitalk
https://api.github.com/repos/moonfox/gitalk
opened
go 语言 基准测试 结果解读 | 睡月花儿
Gitalk go-test-benchmark-result-introducing/
https://www.gagahappy.com/go-test-benchmark-result-introducing/ 基准测试是测量一个程序在固定工作负载下的性能,使用 -bench 标记可以对代码进行基准测试 go test -v -bench=. gott/hello goos: darwin goarch: amd64 pkg: gott/hello BenchmarkHello-4 4964053 228.5 ns/op PASS ok gott/he
1.0
go 语言 基准测试 结果解读 | 睡月花儿 - https://www.gagahappy.com/go-test-benchmark-result-introducing/ 基准测试是测量一个程序在固定工作负载下的性能,使用 -bench 标记可以对代码进行基准测试 go test -v -bench=. gott/hello goos: darwin goarch: amd64 pkg: gott/hello BenchmarkHello-4 4964053 228.5 ns/op PASS ok gott/he
test
go 语言 基准测试 结果解读 睡月花儿 基准测试是测量一个程序在固定工作负载下的性能,使用 bench 标记可以对代码进行基准测试 go test v bench gott hello goos darwin goarch pkg gott hello benchmarkhello ns op pass ok gott he
1
277,168
24,054,126,848
IssuesEvent
2022-09-16 15:13:56
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Certain colors are not reflected in classic themes without theme.json
Needs Testing [Status] Stale Global Styles
### Description In classic themes without `theme.json`, certain colors are not reflected on the frontend. This occurs when a HEX color defined in the core `theme.json` is used. If the editor uses that custom color, the block will be given its own class name. (e.g. `#0693e3` to `vivid-cyan-blue` class name) However, colors are not reflected because inline styles like `.has-vivid-cyan-blue` are not generated for classic themes without `theme.json`. This occurs when the gutenberg plugin is enabled. ### Step-by-step reproduction instructions - Enable classic theme without theme.json ( e.g. Twenty Twenty One). - Change paragraph text and background colors to custom color that is defined as the core color palette. (e.g. Vivid cyan blue: `#0693e3`, Pale pink: `#f78da7`) - Change the editor to code editor mode, and confirm that color class is added to the block attribute. - Save the post, and confirm that the color is not reflected on the frontend. (Additional) - add blank `theme.json` with an empty object to the active theme directory. - Reload the browser, and confirm that the color is reflected on the frontend. ### Screenshots, screen recording, code snippet https://user-images.githubusercontent.com/54422211/160268797-addf32be-9f19-483d-aa61-c0d1eaa2b37f.mp4 ### Environment info - WordPress version: 5.9.2 - Gutenberg version: 12.9.0-rc.1 - Theme: Twenty Twenty One (version1.5) ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
1.0
Certain colors are not reflected in classic themes without theme.json - ### Description In classic themes without `theme.json`, certain colors are not reflected on the frontend. This occurs when a HEX color defined in the core `theme.json` is used. If the editor uses that custom color, the block will be given its own class name. (e.g. `#0693e3` to `vivid-cyan-blue` class name) However, colors are not reflected because inline styles like `.has-vivid-cyan-blue` are not generated for classic themes without `theme.json`. This occurs when the gutenberg plugin is enabled. ### Step-by-step reproduction instructions - Enable classic theme without theme.json ( e.g. Twenty Twenty One). - Change paragraph text and background colors to custom color that is defined as the core color palette. (e.g. Vivid cyan blue: `#0693e3`, Pale pink: `#f78da7`) - Change the editor to code editor mode, and confirm that color class is added to the block attribute. - Save the post, and confirm that the color is not reflected on the frontend. (Additional) - add blank `theme.json` with an empty object to the active theme directory. - Reload the browser, and confirm that the color is reflected on the frontend. ### Screenshots, screen recording, code snippet https://user-images.githubusercontent.com/54422211/160268797-addf32be-9f19-483d-aa61-c0d1eaa2b37f.mp4 ### Environment info - WordPress version: 5.9.2 - Gutenberg version: 12.9.0-rc.1 - Theme: Twenty Twenty One (version1.5) ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
test
certain colors are not reflected in classic themes without theme json description in classic themes without theme json certain colors are not reflected on the frontend this occurs when a hex color defined in the core theme json is used if the editor uses that custom color the block will be given its own class name e g to vivid cyan blue class name however colors are not reflected because inline styles like has vivid cyan blue are not generated for classic themes without theme json this occurs when the gutenberg plugin is enabled step by step reproduction instructions enable classic theme without theme json e g twenty twenty one change paragraph text and background colors to custom color that is defined as the core color palette e g vivid cyan blue pale pink change the editor to code editor mode and confirm that color class is added to the block attribute save the post and confirm that the color is not reflected on the frontend additional add blank theme json with an empty object to the active theme directory reload the browser and confirm that the color is reflected on the frontend screenshots screen recording code snippet environment info wordpress version gutenberg version rc theme twenty twenty one please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
1
240,658
20,067,716,952
IssuesEvent
2022-02-04 00:06:04
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/logstash/node_detail_mb·js - Monitoring app Logstash node detail mb Node "before all" hook for "detail view should have summary status showing correct info"
failed-test test-cloud
**Version: 8.1.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/logstash/node_detail_mb·js** **Stack Trace:** ``` Error: retry.try timeout: Error: expected testSubject(superDatePickerQuickMenu) to exist at TestSubjects.existOrFail (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:44:13) at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:43:11 at runAttempt (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:29:15) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:21) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:59:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at Context.<anonymous> (test/functional/apps/monitoring/logstash/node_detail_mb.js:34:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2876/testReport/_
2.0
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/logstash/node_detail_mb·js - Monitoring app Logstash node detail mb Node "before all" hook for "detail view should have summary status showing correct info" - **Version: 8.1.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/logstash/node_detail_mb·js** **Stack Trace:** ``` Error: retry.try timeout: Error: expected testSubject(superDatePickerQuickMenu) to exist at TestSubjects.existOrFail (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:44:13) at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:43:11 at runAttempt (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:29:15) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:21) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:59:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at Context.<anonymous> (test/functional/apps/monitoring/logstash/node_detail_mb.js:34:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2876/testReport/_
test
chrome x pack ui functional x pack test functional apps monitoring logstash node detail mb·js monitoring app logstash node detail mb node before all hook for detail view should have summary status showing correct info version class chrome x pack ui functional x pack test functional apps monitoring logstash node detail mb·js stack trace error retry try timeout error expected testsubject superdatepickerquickmenu to exist at testsubjects existorfail var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common test subjects ts at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at runattempt var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at setstate var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at object open var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at timepickerpageobject getrefreshconfig var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at timepickerpageobject pauseautorefresh var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at setup test functional apps monitoring get lifecycle methods js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at setstate var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at object open var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at timepickerpageobject getrefreshconfig var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at timepickerpageobject pauseautorefresh var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at setup test functional apps monitoring get lifecycle methods js at context test functional apps monitoring logstash node detail mb js at object apply var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules kbn test target node functional test runner lib mocha wrap function js other test failures test report
1
252,539
19,030,034,114
IssuesEvent
2021-11-24 09:42:29
hirosystems/stacks-blockchain-api
https://api.github.com/repos/hirosystems/stacks-blockchain-api
closed
Missing documentation
documentation
I was looking for documentation for the api client. I found this: https://github.com/hirosystems/stacks-blockchain-api/tree/master/client But the link at the bottom of that page that is supposed to go to the full reference gives a 404.
1.0
Missing documentation - I was looking for documentation for the api client. I found this: https://github.com/hirosystems/stacks-blockchain-api/tree/master/client But the link at the bottom of that page that is supposed to go to the full reference gives a 404.
non_test
missing documentation i was looking for documentation for the api client i found this but the link at the bottom of that page that is supposed to go to the full reference gives a
0
230,189
18,514,502,432
IssuesEvent
2021-10-20 08:36:18
NLCR/SeznamDNNT
https://api.github.com/repos/NLCR/SeznamDNNT
opened
Duplicita vyhledaných záznamů
p:TEST c::Functions p::Middle
Dobrý den, pokud zadám do vyhledávače název dokumentu, např. "Terajší stav slovenskej hudby", pak jsou nalezeny 2 identické záznamy. ![duplicita_vyhledanych_zaznamu](https://user-images.githubusercontent.com/92719629/138058117-37376895-3ba6-4de2-989b-e6f5175b8961.png)
1.0
Duplicita vyhledaných záznamů - Dobrý den, pokud zadám do vyhledávače název dokumentu, např. "Terajší stav slovenskej hudby", pak jsou nalezeny 2 identické záznamy. ![duplicita_vyhledanych_zaznamu](https://user-images.githubusercontent.com/92719629/138058117-37376895-3ba6-4de2-989b-e6f5175b8961.png)
test
duplicita vyhledaných záznamů dobrý den pokud zadám do vyhledávače název dokumentu např terajší stav slovenskej hudby pak jsou nalezeny identické záznamy
1
793,437
27,996,615,626
IssuesEvent
2023-03-27 08:55:20
AY2223S2-CS2103T-W10-2/tp
https://api.github.com/repos/AY2223S2-CS2103T-W10-2/tp
closed
Come up with representative way of handling selection
type.Task priority.High
Currently the edit and delete commands rely on the user to supply an index. We need a way of representing this graphically (i. e. search for tasks and display as a filteredlist).
1.0
Come up with representative way of handling selection - Currently the edit and delete commands rely on the user to supply an index. We need a way of representing this graphically (i. e. search for tasks and display as a filteredlist).
non_test
come up with representative way of handling selection currently the edit and delete commands rely on the user to supply an index we need a way of representing this graphically i e search for tasks and display as a filteredlist
0
705,984
24,256,282,563
IssuesEvent
2022-09-27 18:04:20
dotCMS/core
https://api.github.com/repos/dotCMS/core
closed
Copy folder duplicates content on pages
Type : Bug Passed QA Merged Severity : Support Requested Passed Internal QA LTS: Next Severity : CS Priority Team: Scout Next LTS Release Release : 22.10 Release : 5.3.8.14
**Describe the bug** If you copy a folder that contains a page, the content in the copied page is duplicated. Note: only shows duplicated if the container in the template doesn't have a unique ID. If it has an unique ID it duplicates the content but not showing in UI. **To Reproduce** Steps to reproduce the behavior: 1. Create an Advance Template 2. Click on Add Container and add the Rich Text Container. 3. Delete the unique ID, so it looks like this: `#parseContainer('d71d56b4-0a8b-4bb2-be15-ffa5a23366ea','1')` 4. Save the Template 5. Create a Folder 6. Create a page in the folder with the created template. 7. Add content to the page. 8. Go back to the site browser 9. Right-click over the folder 10. Mark for Copy 11. Right-click over the site 12. Paste 13. Open the `_copy` folder 14. Open the page 15. See the duplicated content Note: check the DB you'll see that the page has 2 content related, one entry shows the `relation_type` as **LEGACY_RELATION_TYPE**. **Expected behavior** Copying folders shouldn't duplicate content on pages. **Workaround** Create the folder manually and only copy the pages. Reproducible in 5.3.8 LTS, 21.06 LTS and demo (22.08). Customer ticket https://dotcms.zendesk.com/agent/tickets/108105
1.0
Copy folder duplicates content on pages - **Describe the bug** If you copy a folder that contains a page, the content in the copied page is duplicated. Note: only shows duplicated if the container in the template doesn't have a unique ID. If it has an unique ID it duplicates the content but not showing in UI. **To Reproduce** Steps to reproduce the behavior: 1. Create an Advance Template 2. Click on Add Container and add the Rich Text Container. 3. Delete the unique ID, so it looks like this: `#parseContainer('d71d56b4-0a8b-4bb2-be15-ffa5a23366ea','1')` 4. Save the Template 5. Create a Folder 6. Create a page in the folder with the created template. 7. Add content to the page. 8. Go back to the site browser 9. Right-click over the folder 10. Mark for Copy 11. Right-click over the site 12. Paste 13. Open the `_copy` folder 14. Open the page 15. See the duplicated content Note: check the DB you'll see that the page has 2 content related, one entry shows the `relation_type` as **LEGACY_RELATION_TYPE**. **Expected behavior** Copying folders shouldn't duplicate content on pages. **Workaround** Create the folder manually and only copy the pages. Reproducible in 5.3.8 LTS, 21.06 LTS and demo (22.08). Customer ticket https://dotcms.zendesk.com/agent/tickets/108105
non_test
copy folder duplicates content on pages describe the bug if you copy a folder that contains a page the content in the copied page is duplicated note only shows duplicated if the container in the template doesn t have a unique id if it has an unique id it duplicates the content but not showing in ui to reproduce steps to reproduce the behavior create an advance template click on add container and add the rich text container delete the unique id so it looks like this parsecontainer save the template create a folder create a page in the folder with the created template add content to the page go back to the site browser right click over the folder mark for copy right click over the site paste open the copy folder open the page see the duplicated content note check the db you ll see that the page has content related one entry shows the relation type as legacy relation type expected behavior copying folders shouldn t duplicate content on pages workaround create the folder manually and only copy the pages reproducible in lts lts and demo customer ticket
0
15,401
9,996,242,591
IssuesEvent
2019-07-11 22:41:01
aegisbigdata/documentation
https://api.github.com/repos/aegisbigdata/documentation
closed
Clarify the results provided by the "Assets" left menu item
USABILITY
Proposal: remove the "Assets" item from the left menu.
True
Clarify the results provided by the "Assets" left menu item - Proposal: remove the "Assets" item from the left menu.
non_test
clarify the results provided by the assets left menu item proposal remove the assets item from the left menu
0
43,627
5,545,823,821
IssuesEvent
2017-03-22 22:39:04
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
closed
Unskip test-cache-sw-core:"fetches new diversions" test
P1: High Priority Related to: Flaky Tests Type: Bug
Test was skipped because it was failing in Chrome 45 with: ``` Chrome 45.0.2454 (Linux 0.0.0) Cache SW fetchJsFile when response is ok when diversions request succeeds with diversions fetches new diversions FAILED expected fetch to have been called exactly thrice, but it was called twice fetch({ url: "https://cdn.ampproject.org/rtv/011234567891234/v0.js" }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 fetch([Request] { }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 AssertionError: expected fetch to have been called exactly thrice, but it was called twice fetch({ url: "https://cdn.ampproject.org/rtv/011234567891234/v0.js" }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 fetch([Request] { }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 at /tmp/f7923860e57dc98f164701657a64065f.browserify:146082:26 <- /home/travis/build/ampproject/amphtml/test/functional/test-cache-sw-core.js:563:26 ```
1.0
Unskip test-cache-sw-core:"fetches new diversions" test - Test was skipped because it was failing in Chrome 45 with: ``` Chrome 45.0.2454 (Linux 0.0.0) Cache SW fetchJsFile when response is ok when diversions request succeeds with diversions fetches new diversions FAILED expected fetch to have been called exactly thrice, but it was called twice fetch({ url: "https://cdn.ampproject.org/rtv/011234567891234/v0.js" }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 fetch([Request] { }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 AssertionError: expected fetch to have been called exactly thrice, but it was called twice fetch({ url: "https://cdn.ampproject.org/rtv/011234567891234/v0.js" }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 fetch([Request] { }) => [Promise] { } at /tmp/f7923860e57dc98f164701657a64065f.browserify:117709:12 <- /home/travis/build/ampproject/amphtml/src/service-worker/core.js:280:15 at /tmp/f7923860e57dc98f164701657a64065f.browserify:146082:26 <- /home/travis/build/ampproject/amphtml/test/functional/test-cache-sw-core.js:563:26 ```
test
unskip test cache sw core fetches new diversions test test was skipped because it was failing in chrome with chrome linux cache sw fetchjsfile when response is ok when diversions request succeeds with diversions fetches new diversions failed expected fetch to have been called exactly thrice but it was called twice fetch url at tmp browserify home travis build ampproject amphtml src service worker core js fetch at tmp browserify home travis build ampproject amphtml src service worker core js assertionerror expected fetch to have been called exactly thrice but it was called twice fetch url at tmp browserify home travis build ampproject amphtml src service worker core js fetch at tmp browserify home travis build ampproject amphtml src service worker core js at tmp browserify home travis build ampproject amphtml test functional test cache sw core js
1
281,651
24,410,488,505
IssuesEvent
2022-10-05 11:56:50
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts - saved objects security and spaces enabled _bulk_create user with no access within the default space "before all" hook for "should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id]"
blocker Team:Security failed-test skipped-test v8.4.0
A test failed on a tracked branch ``` ResponseError: illegal_argument_exception: [illegal_argument_exception] Reason: request [/.kibana_7.17.0_001] contains unrecognized parameter: [include_type_name] at onBody (/opt/local-ssd/buildkite/builds/kb-n2-4-bf1ddb6b4a216c51/elastic/kibana-7-dot-latest-es-forward-compatibility/kibana/node_modules/@elastic/elasticsearch/lib/Transport.js:367:23) at IncomingMessage.onEnd (/opt/local-ssd/buildkite/builds/kb-n2-4-bf1ddb6b4a216c51/elastic/kibana-7-dot-latest-es-forward-compatibility/kibana/node_modules/@elastic/elasticsearch/lib/Transport.js:291:11) at IncomingMessage.emit (node:events:402:35) at endReadableNT (node:internal/streams/readable:1343:12) at processTicksAndRejections (node:internal/process/task_queues:83:21) { meta: { body: { error: [Object], status: 400 }, statusCode: 400, headers: { 'x-elastic-product': 'Elasticsearch', 'content-type': 'application/json;charset=utf-8', 'content-length': '301' }, meta: { context: null, request: [Object], name: 'elasticsearch-js', connection: [Object], attempts: 0, aborted: false } } } ``` First failure: [CI Build - 7.17](https://buildkite.com/elastic/kibana-7-dot-latest-es-forward-compatibility/builds/1#e408c0a1-ecf6-4057-87fb-386a95c671b3) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts","test.name":"saved objects security and spaces enabled _bulk_create user with no access within the default space \"before all\" hook for \"should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id]\"","test.failCount":5}} -->
2.0
Failing test: X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts - saved objects security and spaces enabled _bulk_create user with no access within the default space "before all" hook for "should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id]" - A test failed on a tracked branch ``` ResponseError: illegal_argument_exception: [illegal_argument_exception] Reason: request [/.kibana_7.17.0_001] contains unrecognized parameter: [include_type_name] at onBody (/opt/local-ssd/buildkite/builds/kb-n2-4-bf1ddb6b4a216c51/elastic/kibana-7-dot-latest-es-forward-compatibility/kibana/node_modules/@elastic/elasticsearch/lib/Transport.js:367:23) at IncomingMessage.onEnd (/opt/local-ssd/buildkite/builds/kb-n2-4-bf1ddb6b4a216c51/elastic/kibana-7-dot-latest-es-forward-compatibility/kibana/node_modules/@elastic/elasticsearch/lib/Transport.js:291:11) at IncomingMessage.emit (node:events:402:35) at endReadableNT (node:internal/streams/readable:1343:12) at processTicksAndRejections (node:internal/process/task_queues:83:21) { meta: { body: { error: [Object], status: 400 }, statusCode: 400, headers: { 'x-elastic-product': 'Elasticsearch', 'content-type': 'application/json;charset=utf-8', 'content-length': '301' }, meta: { context: null, request: [Object], name: 'elasticsearch-js', connection: [Object], attempts: 0, aborted: false } } } ``` First failure: [CI Build - 7.17](https://buildkite.com/elastic/kibana-7-dot-latest-es-forward-compatibility/builds/1#e408c0a1-ecf6-4057-87fb-386a95c671b3) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts","test.name":"saved objects security and spaces enabled _bulk_create user with no access within the default space \"before all\" hook for \"should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id]\"","test.failCount":5}} -->
test
failing test x pack saved object api integration tests security and spaces x pack test saved object api integration security and spaces apis bulk create·ts saved objects security and spaces enabled bulk create user with no access within the default space before all hook for should return forbidden a test failed on a tracked branch responseerror illegal argument exception reason request contains unrecognized parameter at onbody opt local ssd buildkite builds kb elastic kibana dot latest es forward compatibility kibana node modules elastic elasticsearch lib transport js at incomingmessage onend opt local ssd buildkite builds kb elastic kibana dot latest es forward compatibility kibana node modules elastic elasticsearch lib transport js at incomingmessage emit node events at endreadablent node internal streams readable at processticksandrejections node internal process task queues meta body error status statuscode headers x elastic product elasticsearch content type application json charset utf content length meta context null request name elasticsearch js connection attempts aborted false first failure
1
86,303
15,755,515,348
IssuesEvent
2021-03-31 01:55:09
biswajit-paul/dpone
https://api.github.com/repos/biswajit-paul/dpone
opened
CVE-2020-28500 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.11.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-0.9.2.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/form-data/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-qunit-0.4.0.tgz (Root Library) - grunt-lib-phantomjs-0.5.0.tgz - phantomjs-1.9.20.tgz - request-2.67.0.tgz - form-data-1.0.1.tgz - async-2.6.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/grunt-bowercopy/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.2.tgz (Root Library) - findup-sync-0.1.3.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.9.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz">https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.2.tgz (Root Library) - :x: **lodash-0.9.2.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7">https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28500 (Medium) detected in multiple libraries - ## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.11.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-0.9.2.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/form-data/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-qunit-0.4.0.tgz (Root Library) - grunt-lib-phantomjs-0.5.0.tgz - phantomjs-1.9.20.tgz - request-2.67.0.tgz - form-data-1.0.1.tgz - async-2.6.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/grunt-bowercopy/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.2.tgz (Root Library) - findup-sync-0.1.3.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.9.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz">https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz</a></p> <p>Path to dependency file: /dpone/core/assets/vendor/jquery.ui/package.json</p> <p>Path to vulnerable library: dpone/core/assets/vendor/jquery.ui/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.2.tgz (Root Library) - :x: **lodash-0.9.2.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7">https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file dpone core assets vendor jquery ui package json path to vulnerable library dpone core assets vendor jquery ui node modules form data node modules lodash package json dependency hierarchy grunt contrib qunit tgz root library grunt lib phantomjs tgz phantomjs tgz request tgz form data tgz async tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file dpone core assets vendor jquery ui package json path to vulnerable library dpone core assets vendor jquery ui node modules grunt bowercopy node modules lodash package json dependency hierarchy grunt tgz root library findup sync tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file dpone core assets vendor jquery ui package json path to vulnerable library dpone core assets vendor jquery ui node modules lodash package json dependency hierarchy grunt tgz root library x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
332,623
29,490,044,812
IssuesEvent
2023-06-02 12:50:12
CDCgov/data-exchange-upload
https://api.github.com/repos/CDCgov/data-exchange-upload
reopened
Upload API stress/load testing
testing
### Stress test the following scenarios: - [x] Multiple clients with multiple simultaneous uploads (one upload per client) - [x] One client with multiple simultaneous uploads under same client session - [ ] Single large file upload -- this is covered by issue #85 - [x] Simultaneous large file uploads --- - [x] Test DEX -> EDAV copy by azure function for large payloads Note: please see test results below.
1.0
Upload API stress/load testing - ### Stress test the following scenarios: - [x] Multiple clients with multiple simultaneous uploads (one upload per client) - [x] One client with multiple simultaneous uploads under same client session - [ ] Single large file upload -- this is covered by issue #85 - [x] Simultaneous large file uploads --- - [x] Test DEX -> EDAV copy by azure function for large payloads Note: please see test results below.
test
upload api stress load testing stress test the following scenarios multiple clients with multiple simultaneous uploads one upload per client one client with multiple simultaneous uploads under same client session single large file upload this is covered by issue simultaneous large file uploads test dex edav copy by azure function for large payloads note please see test results below
1
262,518
22,909,093,011
IssuesEvent
2022-07-16 02:34:51
SEED-platform/seed
https://api.github.com/repos/SEED-platform/seed
closed
ESPM Data Import Path
Testing Question Include Before Closure Requires comms after completion
ESPM provides two ways to generate buidling data: **Download My Portfolio** - currently imports via "Upload a Spreadsheet" with no issues - works if the program participants “share their property” with you in ESPM **Data Request** - imports via "Upload a Spreadsheet" with "Import Meter from the Same File" resulting in anonymous errors - some file manipulation allows importing meters, but data is restricted based on meter types and does funny stuff (see request below) - imports via "Import Portfolio Manager Data" result in no meters or errors. Users would like to be able to import properties and meters using the "Data Request" path. These changes are requested: - provide a more detailed error when importing meters so users know what to do - "This is where I’m not sure if we can upload a file that has more than just the Elec and Gas meter – and we would like to be able to upload multiple meters, including district heating energy, propane, etc." - "Also, if we are just uploading Elec and Nat Gas, the elec meter gets named “Electric – Unknown – PM – Unknown” and gas gets named “Natural Gas – PM – Unknown”. Contact: Open Tech
1.0
ESPM Data Import Path - ESPM provides two ways to generate buidling data: **Download My Portfolio** - currently imports via "Upload a Spreadsheet" with no issues - works if the program participants “share their property” with you in ESPM **Data Request** - imports via "Upload a Spreadsheet" with "Import Meter from the Same File" resulting in anonymous errors - some file manipulation allows importing meters, but data is restricted based on meter types and does funny stuff (see request below) - imports via "Import Portfolio Manager Data" result in no meters or errors. Users would like to be able to import properties and meters using the "Data Request" path. These changes are requested: - provide a more detailed error when importing meters so users know what to do - "This is where I’m not sure if we can upload a file that has more than just the Elec and Gas meter – and we would like to be able to upload multiple meters, including district heating energy, propane, etc." - "Also, if we are just uploading Elec and Nat Gas, the elec meter gets named “Electric – Unknown – PM – Unknown” and gas gets named “Natural Gas – PM – Unknown”. Contact: Open Tech
test
espm data import path espm provides two ways to generate buidling data download my portfolio currently imports via upload a spreadsheet with no issues works if the program participants “share their property” with you in espm data request imports via upload a spreadsheet with import meter from the same file resulting in anonymous errors some file manipulation allows importing meters but data is restricted based on meter types and does funny stuff see request below imports via import portfolio manager data result in no meters or errors users would like to be able to import properties and meters using the data request path these changes are requested provide a more detailed error when importing meters so users know what to do this is where i’m not sure if we can upload a file that has more than just the elec and gas meter – and we would like to be able to upload multiple meters including district heating energy propane etc also if we are just uploading elec and nat gas the elec meter gets named “electric – unknown – pm – unknown” and gas gets named “natural gas – pm – unknown” contact open tech
1
330,688
28,481,932,088
IssuesEvent
2023-04-18 03:57:11
wpfoodmanager/wp-food-manager
https://api.github.com/repos/wpfoodmanager/wp-food-manager
closed
Admin : Field editor tab group title should be change
In Testing Issue Resolved
Admin side Field editor tab group title should be change it is working only in admin side so need to mention this. ![image](https://user-images.githubusercontent.com/121149500/232445134-19bc8f72-7e4a-48a7-9a10-5ccbbc018994.png)
1.0
Admin : Field editor tab group title should be change - Admin side Field editor tab group title should be change it is working only in admin side so need to mention this. ![image](https://user-images.githubusercontent.com/121149500/232445134-19bc8f72-7e4a-48a7-9a10-5ccbbc018994.png)
test
admin field editor tab group title should be change admin side field editor tab group title should be change it is working only in admin side so need to mention this
1
115,079
11,864,915,088
IssuesEvent
2020-03-25 22:49:25
JAD3N/java-props
https://api.github.com/repos/JAD3N/java-props
closed
Add README.md
documentation
Need to add a README.md file for future reference to understand the usage of the crate.
1.0
Add README.md - Need to add a README.md file for future reference to understand the usage of the crate.
non_test
add readme md need to add a readme md file for future reference to understand the usage of the crate
0
31,255
4,699,813,480
IssuesEvent
2016-10-12 16:43:24
MachoThemes/newsmag-lite
https://api.github.com/repos/MachoThemes/newsmag-lite
closed
Improve w.org theme demo
tested
Trebuie sa adaugam demo data din widgetzii default ai WP-ului pe sidebar-ul de blog si pe sidebar-urile din footer.
1.0
Improve w.org theme demo - Trebuie sa adaugam demo data din widgetzii default ai WP-ului pe sidebar-ul de blog si pe sidebar-urile din footer.
test
improve w org theme demo trebuie sa adaugam demo data din widgetzii default ai wp ului pe sidebar ul de blog si pe sidebar urile din footer
1
439,249
12,680,092,338
IssuesEvent
2020-06-19 13:05:38
naev/naev
https://api.github.com/repos/naev/naev
closed
German translation
Priority-Low Type-Enhancement
Hello! I just post this issue to inform you that I have today started to work on adding a German translation. I am just giving you a heads-up, to avoid potential clashes. I will talk back to you when I am finished.
1.0
German translation - Hello! I just post this issue to inform you that I have today started to work on adding a German translation. I am just giving you a heads-up, to avoid potential clashes. I will talk back to you when I am finished.
non_test
german translation hello i just post this issue to inform you that i have today started to work on adding a german translation i am just giving you a heads up to avoid potential clashes i will talk back to you when i am finished
0
288,468
24,905,746,216
IssuesEvent
2022-10-29 08:05:15
f2etw/jobs
https://api.github.com/repos/f2etw/jobs
closed
[DeFi Web3 FinTech 全遠端 徵才][Aha Technologies] Sr Front & Backend Engineer | 資深 前/後端工程師 年薪3百萬
求才 senior [F] React Full Stack unit test 海外 Work Remotely
**| 關於我們** 本司是一家來自矽谷的 FinTech新創軟體公司,由來自Stanford 史丹佛和 MIT 麻省理工學院的校友共同創立。我們的創辦人、投資者與顧問團隊皆擁有科技業、金融和管理諮詢的豐富經歷,致力於創造非凡的軟體解決全球用戶的問題! 是否非常嚮往不受時間與空間拘束的企業制度? 沒錯!如果你: 渴望自己決定工時,彈性上下班的制度 渴望遠端自由的環境,想逃離辦公室 渴望有挑戰、發展和前景的新創領域 渴望在團隊中擁有參與感和主導權 具有創業和冒險精神,不害怕未知的挑戰 對國際市場有熱情和野心,未來想到美國工作 你就是我們在尋找的 [資深] 前/後端網頁開發工程師!希望我們的新創團隊成為你理想與事業啟航的起點! https://earnaha.com/ **【申請流程】** 有興趣請填寫此申請表,上傳履歷表,並預約線上面談: https://linktr.ee/growaha **【前端職務內容】** 規劃及開發全端系統/平台的核心邏輯 **【前端經驗條件】** - 資訊、電腦工程、電機,相關學科 - 五年以上的程式學科或工作實習經驗 - 現代全端開發 **JavaScript** (Node, Express, React, Redux)、Django、Rails、HTML、CSS 經驗 - 數據庫經驗(如 PostgreSQL,MySQL,MongoDB) - 測試QC經驗(如 QUnit,Mocha,Jasmine,RSpec) - 虛擬化經驗(如 Docker,K8S) **【後端職務內容】** 規劃及開發後端系統/平台的核心邏輯,並與前端合作,串接 APIs **【後端經驗條件】** - 資訊,電腦工程,電機,相關學科 - 五年以上的程式學科或工作實習經驗 - 網站框架經驗(列Ruby on Rails,Django-Python,Java,Node.js) - 數據庫經驗(列PostgreSQL,SQL Server) - 測試QC經驗(列QUnit,Mocha,Jasmine,RSpec) **【加分條件】** - 自我激勵、自我學習、自我管理、自我要求 - 渴望滿足使用者的需求,並追求成就與卓越感 - 能夠快速學習新技能和熟悉新領域 - 願意主動參與團隊,與非工程人員一同合作奮鬥 - 擁有管理經驗 **【福利待遇】** 1. 可配合個人職涯規劃,自由選擇長期或短期工作 2. 時間彈性、遠端工作(無硬性規定上班時間,也無須進辦公室,只要按時完成份內工作即可。) 3. 績效獎金 **【薪資等級】** 新員工的職位等級由考試決定:https://linktr.ee/growaha **【公司詳細資訊】** 投履歷與面試前,請先閱讀: https://linktr.ee/growaha **【申請流程】** 有興趣請填寫此申請表,上傳履歷表,並預約線上面談: https://linktr.ee/growaha
1.0
[DeFi Web3 FinTech 全遠端 徵才][Aha Technologies] Sr Front & Backend Engineer | 資深 前/後端工程師 年薪3百萬 - **| 關於我們** 本司是一家來自矽谷的 FinTech新創軟體公司,由來自Stanford 史丹佛和 MIT 麻省理工學院的校友共同創立。我們的創辦人、投資者與顧問團隊皆擁有科技業、金融和管理諮詢的豐富經歷,致力於創造非凡的軟體解決全球用戶的問題! 是否非常嚮往不受時間與空間拘束的企業制度? 沒錯!如果你: 渴望自己決定工時,彈性上下班的制度 渴望遠端自由的環境,想逃離辦公室 渴望有挑戰、發展和前景的新創領域 渴望在團隊中擁有參與感和主導權 具有創業和冒險精神,不害怕未知的挑戰 對國際市場有熱情和野心,未來想到美國工作 你就是我們在尋找的 [資深] 前/後端網頁開發工程師!希望我們的新創團隊成為你理想與事業啟航的起點! https://earnaha.com/ **【申請流程】** 有興趣請填寫此申請表,上傳履歷表,並預約線上面談: https://linktr.ee/growaha **【前端職務內容】** 規劃及開發全端系統/平台的核心邏輯 **【前端經驗條件】** - 資訊、電腦工程、電機,相關學科 - 五年以上的程式學科或工作實習經驗 - 現代全端開發 **JavaScript** (Node, Express, React, Redux)、Django、Rails、HTML、CSS 經驗 - 數據庫經驗(如 PostgreSQL,MySQL,MongoDB) - 測試QC經驗(如 QUnit,Mocha,Jasmine,RSpec) - 虛擬化經驗(如 Docker,K8S) **【後端職務內容】** 規劃及開發後端系統/平台的核心邏輯,並與前端合作,串接 APIs **【後端經驗條件】** - 資訊,電腦工程,電機,相關學科 - 五年以上的程式學科或工作實習經驗 - 網站框架經驗(列Ruby on Rails,Django-Python,Java,Node.js) - 數據庫經驗(列PostgreSQL,SQL Server) - 測試QC經驗(列QUnit,Mocha,Jasmine,RSpec) **【加分條件】** - 自我激勵、自我學習、自我管理、自我要求 - 渴望滿足使用者的需求,並追求成就與卓越感 - 能夠快速學習新技能和熟悉新領域 - 願意主動參與團隊,與非工程人員一同合作奮鬥 - 擁有管理經驗 **【福利待遇】** 1. 可配合個人職涯規劃,自由選擇長期或短期工作 2. 時間彈性、遠端工作(無硬性規定上班時間,也無須進辦公室,只要按時完成份內工作即可。) 3. 績效獎金 **【薪資等級】** 新員工的職位等級由考試決定:https://linktr.ee/growaha **【公司詳細資訊】** 投履歷與面試前,請先閱讀: https://linktr.ee/growaha **【申請流程】** 有興趣請填寫此申請表,上傳履歷表,並預約線上面談: https://linktr.ee/growaha
test
sr front backend engineer 資深 前 後端工程師 關於我們 本司是一家來自矽谷的 fintech新創軟體公司,由來自stanford 史丹佛和 mit 麻省理工學院的校友共同創立。我們的創辦人、投資者與顧問團隊皆擁有科技業、金融和管理諮詢的豐富經歷,致力於創造非凡的軟體解決全球用戶的問題! 是否非常嚮往不受時間與空間拘束的企業制度? 沒錯!如果你: 渴望自己決定工時,彈性上下班的制度 渴望遠端自由的環境,想逃離辦公室 渴望有挑戰、發展和前景的新創領域 渴望在團隊中擁有參與感和主導權 具有創業和冒險精神,不害怕未知的挑戰 對國際市場有熱情和野心,未來想到美國工作 你就是我們在尋找的 前 後端網頁開發工程師!希望我們的新創團隊成為你理想與事業啟航的起點! 【申請流程】 有興趣請填寫此申請表,上傳履歷表,並預約線上面談 【前端職務內容】 規劃及開發全端系統 平台的核心邏輯 【前端經驗條件】 資訊、電腦工程、電機,相關學科 五年以上的程式學科或工作實習經驗 現代全端開發 javascript node express react redux 、django、rails、html、css 經驗 數據庫經驗(如 postgresql,mysql,mongodb) 測試qc經驗(如 qunit,mocha,jasmine,rspec) 虛擬化經驗(如 docker, ) 【後端職務內容】 規劃及開發後端系統 平台的核心邏輯,並與前端合作,串接 apis 【後端經驗條件】 資訊,電腦工程,電機,相關學科 五年以上的程式學科或工作實習經驗 網站框架經驗(列ruby on rails,django python,java,node js) 數據庫經驗(列postgresql,sql server) 測試qc經驗(列qunit,mocha,jasmine,rspec) 【加分條件】 自我激勵、自我學習、自我管理、自我要求 渴望滿足使用者的需求,並追求成就與卓越感 能夠快速學習新技能和熟悉新領域 願意主動參與團隊,與非工程人員一同合作奮鬥 擁有管理經驗 【福利待遇】 可配合個人職涯規劃,自由選擇長期或短期工作 時間彈性、遠端工作(無硬性規定上班時間,也無須進辦公室,只要按時完成份內工作即可。) 績效獎金 【薪資等級】 新員工的職位等級由考試決定: 【公司詳細資訊】 投履歷與面試前,請先閱讀 【申請流程】 有興趣請填寫此申請表,上傳履歷表,並預約線上面談
1
28,148
2,700,035,022
IssuesEvent
2015-04-03 21:51:49
laravel/framework
https://api.github.com/repos/laravel/framework
closed
forceCreate no longer works
bug database high priority
See pull request #8213 for the bug description ``` User::forceCreate(array( 'name' => 'Steven' 'email' => 'steven@example.org' )); ``` Error is: Cannot instantiate abstract class Model
1.0
forceCreate no longer works - See pull request #8213 for the bug description ``` User::forceCreate(array( 'name' => 'Steven' 'email' => 'steven@example.org' )); ``` Error is: Cannot instantiate abstract class Model
non_test
forcecreate no longer works see pull request for the bug description user forcecreate array name steven email steven example org error is cannot instantiate abstract class model
0
288,834
24,940,116,316
IssuesEvent
2022-10-31 18:11:50
elyra-ai/elyra
https://api.github.com/repos/elyra-ai/elyra
opened
Investigate CI tests execution order
component:test kind:investigative component:script-debugger
**Describe the issue** Follow up issue to problems found in #2971 with respect to the execution order of integration test files. In the PR above a new test file was created containing tests for the script debugger, although all tests would fail remotely, unless the new test file runs first. Investigate which test file might be affecting the script debugger tests to fail remotely and why. **Expected behavior** CI tests should run independent of each other. Make sure the test environment is always cleaned up and reset at the end of tests. The integration test files should be able to run in alphabetical order as expected. **Context information** - Elyra version: v3.13.0-dev - Installation source: src - Operating system: macos
1.0
Investigate CI tests execution order - **Describe the issue** Follow up issue to problems found in #2971 with respect to the execution order of integration test files. In the PR above a new test file was created containing tests for the script debugger, although all tests would fail remotely, unless the new test file runs first. Investigate which test file might be affecting the script debugger tests to fail remotely and why. **Expected behavior** CI tests should run independent of each other. Make sure the test environment is always cleaned up and reset at the end of tests. The integration test files should be able to run in alphabetical order as expected. **Context information** - Elyra version: v3.13.0-dev - Installation source: src - Operating system: macos
test
investigate ci tests execution order describe the issue follow up issue to problems found in with respect to the execution order of integration test files in the pr above a new test file was created containing tests for the script debugger although all tests would fail remotely unless the new test file runs first investigate which test file might be affecting the script debugger tests to fail remotely and why expected behavior ci tests should run independent of each other make sure the test environment is always cleaned up and reset at the end of tests the integration test files should be able to run in alphabetical order as expected context information elyra version dev installation source src operating system macos
1
342,458
30,623,158,167
IssuesEvent
2023-07-24 09:38:09
Qiskit/qiskit.org
https://api.github.com/repos/Qiskit/qiskit.org
closed
Add unit test for hooks
testing
### Feature description ## Why Hooks are our connection to the backend and the is some logic in them that needs to be tested to ensure that the edge cases work propertly. ## Tools to implement them Since vitest is already part of the project, it will be the perfect tool to develop the unit tests. ### Additional context _No response_
1.0
Add unit test for hooks - ### Feature description ## Why Hooks are our connection to the backend and the is some logic in them that needs to be tested to ensure that the edge cases work propertly. ## Tools to implement them Since vitest is already part of the project, it will be the perfect tool to develop the unit tests. ### Additional context _No response_
test
add unit test for hooks feature description why hooks are our connection to the backend and the is some logic in them that needs to be tested to ensure that the edge cases work propertly tools to implement them since vitest is already part of the project it will be the perfect tool to develop the unit tests additional context no response
1
329,099
28,149,844,796
IssuesEvent
2023-04-02 22:29:57
TaleStation/TaleStation
https://api.github.com/repos/TaleStation/TaleStation
closed
Flaky test create_and_destroy: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583)
🤖 Flaky Test Report
<!-- This issue can be renamed, but do not change the next comment! --> <!-- title: Flaky test create_and_destroy: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583) --> Flaky tests were detected in [this test run](https://github.com/TaleStation/TaleStation/actions/runs/4461255773/attempts/1). This means that there was a failure that was cleared when the tests were simply restarted. Failures: ``` create_and_destroy: [15:11:01] Runtime in stack_trace.dm,4: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583) proc name: stack trace (/proc/_stack_trace) src: null call stack: stack trace("addtimer called with a callbac...", "code/controllers/subsystem/tim...", 583) addtimer(/datum/callback (/datum/callback), 300, 8, null, "code/modules/mob/living/simple...", 595) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): GainPatience() the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): GiveTarget(the mi-go (/mob/living/basic/migo)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): FindTarget(/list (/list)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): AIShouldSleep(/list (/list)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): handle automated action() at stack_trace.dm:4 ```
1.0
Flaky test create_and_destroy: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583) - <!-- This issue can be renamed, but do not change the next comment! --> <!-- title: Flaky test create_and_destroy: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583) --> Flaky tests were detected in [this test run](https://github.com/TaleStation/TaleStation/actions/runs/4461255773/attempts/1). This means that there was a failure that was cleared when the tests were simply restarted. Failures: ``` create_and_destroy: [15:11:01] Runtime in stack_trace.dm,4: addtimer called with a callback assigned to a qdeleted object. In the future such timers will not be supported and may refuse to run or run with a 0 wait (code/controllers/subsystem/timer.dm:583) proc name: stack trace (/proc/_stack_trace) src: null call stack: stack trace("addtimer called with a callbac...", "code/controllers/subsystem/tim...", 583) addtimer(/datum/callback (/datum/callback), 300, 8, null, "code/modules/mob/living/simple...", 595) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): GainPatience() the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): GiveTarget(the mi-go (/mob/living/basic/migo)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): FindTarget(/list (/list)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): AIShouldSleep(/list (/list)) the demonic watcher (/mob/living/simple_animal/hostile/asteroid/ice_demon): handle automated action() at stack_trace.dm:4 ```
test
flaky test create and destroy addtimer called with a callback assigned to a qdeleted object in the future such timers will not be supported and may refuse to run or run with a wait code controllers subsystem timer dm flaky tests were detected in this means that there was a failure that was cleared when the tests were simply restarted failures create and destroy runtime in stack trace dm addtimer called with a callback assigned to a qdeleted object in the future such timers will not be supported and may refuse to run or run with a wait code controllers subsystem timer dm proc name stack trace proc stack trace src null call stack stack trace addtimer called with a callbac code controllers subsystem tim addtimer datum callback datum callback null code modules mob living simple the demonic watcher mob living simple animal hostile asteroid ice demon gainpatience the demonic watcher mob living simple animal hostile asteroid ice demon givetarget the mi go mob living basic migo the demonic watcher mob living simple animal hostile asteroid ice demon findtarget list list the demonic watcher mob living simple animal hostile asteroid ice demon aishouldsleep list list the demonic watcher mob living simple animal hostile asteroid ice demon handle automated action at stack trace dm
1
104,914
13,133,766,305
IssuesEvent
2020-08-06 21:37:23
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
opened
Trigger receipt of completed IHPs (VLJ support notification)
Priority: Medium Stakeholder: BVA Team: Echo 🐬 Type: New Development Type: design 💅
<!-- The goal of this template is to be a tool to help write good design, research, and writing tasks that stem off of a user story and/or epic. It is not intended as a mandate, adapt as needed. --> Ticket related to [#13800](https://app.zenhub.com/workspaces/caseflow-5915dd178f67e20b5553ba0c/issues/department-of-veterans-affairs/caseflow/13800) ## User story As a VLJ support staff member I need to be able to see a notification or indicator when an IHP task is completed by a VSO so that I can audit the document and ensure it is complete for a judge/attorney to review. ## Problem statement Per the Board, they do not know when an IHP task is completed in Caseflow and VLJ support plays a key role as they are the bridge between the VSOs and the judges/attorneys. ## What is out of scope? There should be no need to change task assignments. The request here is just for a notification. ## Background/context IHPs are documents that are filled out by VSOs that include additional information related to veteran's case. When a veteran has received assistance from VSO to process their appeal, an IHP is added to the case. At this time VSOs will either complete IHP tasks soon after a veteran has used their services, or they will receive an IHP tasks from a judge or an attorney for a Veterans case when it is being reviewed by a judge/attorney. Once the task is marked as completed there is no notification or indication in Caseflow shown to the Board members (attorneys & judges) causing delays in case movement. ## What are the unknowns? Are cases with open IHP tasks put in the "On-Hold" page in judge/attorney queues and then moved out to "Assigned" once completed? Yes, they do. Could the "Days Waiting" field be some sort of indicator? ## For research tickets only: What questions do we hope to answer? What are our learning goals? How often are completed IHP tasks overlooked? How long of a delay is there between the completion of the IHP task and the attorney or judge picking the case back up? Is there any reason why they would not want to see an alert or indicator for completed IHP tasks? (ex. there is another blocking task against the case so the case cannot be worked in any case) Is there any data we can pull that can possibly decrease the priority of this request? For example, if we find attorneys are picking up these appeals within days of the IHP task being completed is a "notification" really going to make a significant impact? ## For writing tickets only: Writing guidance, if possible, such as voice and tone desired, relevant styleguides, and who the approvers should be. ## Existing design and content [Mural](https://app.mural.co/t/workqueue2001/m/workqueue2001/1595518321333/a3929ce45595dca9177fe2adfd3cb3d6e3203fbd) of research and flows ## Success criteria <!-- Include as needed, especially for issues that aren't part of epics. if no measurable success criteria, what does success look like? --> ## Technical/logistical constraints (if known) <!-- Are there technical constraints that will impact any design or writing solution? Logistical constraints that will impact user research? -->
1.0
Trigger receipt of completed IHPs (VLJ support notification) - <!-- The goal of this template is to be a tool to help write good design, research, and writing tasks that stem off of a user story and/or epic. It is not intended as a mandate, adapt as needed. --> Ticket related to [#13800](https://app.zenhub.com/workspaces/caseflow-5915dd178f67e20b5553ba0c/issues/department-of-veterans-affairs/caseflow/13800) ## User story As a VLJ support staff member I need to be able to see a notification or indicator when an IHP task is completed by a VSO so that I can audit the document and ensure it is complete for a judge/attorney to review. ## Problem statement Per the Board, they do not know when an IHP task is completed in Caseflow and VLJ support plays a key role as they are the bridge between the VSOs and the judges/attorneys. ## What is out of scope? There should be no need to change task assignments. The request here is just for a notification. ## Background/context IHPs are documents that are filled out by VSOs that include additional information related to veteran's case. When a veteran has received assistance from VSO to process their appeal, an IHP is added to the case. At this time VSOs will either complete IHP tasks soon after a veteran has used their services, or they will receive an IHP tasks from a judge or an attorney for a Veterans case when it is being reviewed by a judge/attorney. Once the task is marked as completed there is no notification or indication in Caseflow shown to the Board members (attorneys & judges) causing delays in case movement. ## What are the unknowns? Are cases with open IHP tasks put in the "On-Hold" page in judge/attorney queues and then moved out to "Assigned" once completed? Yes, they do. Could the "Days Waiting" field be some sort of indicator? ## For research tickets only: What questions do we hope to answer? What are our learning goals? How often are completed IHP tasks overlooked? How long of a delay is there between the completion of the IHP task and the attorney or judge picking the case back up? Is there any reason why they would not want to see an alert or indicator for completed IHP tasks? (ex. there is another blocking task against the case so the case cannot be worked in any case) Is there any data we can pull that can possibly decrease the priority of this request? For example, if we find attorneys are picking up these appeals within days of the IHP task being completed is a "notification" really going to make a significant impact? ## For writing tickets only: Writing guidance, if possible, such as voice and tone desired, relevant styleguides, and who the approvers should be. ## Existing design and content [Mural](https://app.mural.co/t/workqueue2001/m/workqueue2001/1595518321333/a3929ce45595dca9177fe2adfd3cb3d6e3203fbd) of research and flows ## Success criteria <!-- Include as needed, especially for issues that aren't part of epics. if no measurable success criteria, what does success look like? --> ## Technical/logistical constraints (if known) <!-- Are there technical constraints that will impact any design or writing solution? Logistical constraints that will impact user research? -->
non_test
trigger receipt of completed ihps vlj support notification ticket related to user story as a vlj support staff member i need to be able to see a notification or indicator when an ihp task is completed by a vso so that i can audit the document and ensure it is complete for a judge attorney to review problem statement per the board they do not know when an ihp task is completed in caseflow and vlj support plays a key role as they are the bridge between the vsos and the judges attorneys what is out of scope there should be no need to change task assignments the request here is just for a notification background context ihps are documents that are filled out by vsos that include additional information related to veteran s case when a veteran has received assistance from vso to process their appeal an ihp is added to the case at this time vsos will either complete ihp tasks soon after a veteran has used their services or they will receive an ihp tasks from a judge or an attorney for a veterans case when it is being reviewed by a judge attorney once the task is marked as completed there is no notification or indication in caseflow shown to the board members attorneys judges causing delays in case movement what are the unknowns are cases with open ihp tasks put in the on hold page in judge attorney queues and then moved out to assigned once completed yes they do could the days waiting field be some sort of indicator for research tickets only what questions do we hope to answer what are our learning goals how often are completed ihp tasks overlooked how long of a delay is there between the completion of the ihp task and the attorney or judge picking the case back up is there any reason why they would not want to see an alert or indicator for completed ihp tasks ex there is another blocking task against the case so the case cannot be worked in any case is there any data we can pull that can possibly decrease the priority of this request for example if we find attorneys are picking up these appeals within days of the ihp task being completed is a notification really going to make a significant impact for writing tickets only writing guidance if possible such as voice and tone desired relevant styleguides and who the approvers should be existing design and content of research and flows success criteria technical logistical constraints if known
0
155,406
19,802,840,360
IssuesEvent
2022-01-19 01:04:01
maorkuriel/ksa
https://api.github.com/repos/maorkuriel/ksa
opened
CVE-2022-23305 (High) detected in log4j-1.2.16.jar
security vulnerability
## CVE-2022-23305 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.16.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Library home page: <a href="http://logging.apache.org/log4j/1.2/">http://logging.apache.org/log4j/1.2/</a></p> <p>Path to vulnerable library: /ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/log4j-1.2.16.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.16.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2022-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23305>CVE-2022-23305</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.16","packageFilePaths":[null],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.16","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23305","vulnerabilityDetails":"By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23305","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-23305 (High) detected in log4j-1.2.16.jar - ## CVE-2022-23305 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.16.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Library home page: <a href="http://logging.apache.org/log4j/1.2/">http://logging.apache.org/log4j/1.2/</a></p> <p>Path to vulnerable library: /ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/log4j-1.2.16.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.16.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2022-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23305>CVE-2022-23305</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.16","packageFilePaths":[null],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.16","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23305","vulnerabilityDetails":"By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23305","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in jar cve high severity vulnerability vulnerable library jar apache library home page a href path to vulnerable library ksa web root ksa web target root web inf lib jar dependency hierarchy x jar vulnerable library found in base branch master vulnerability details by design the jdbcappender in x accepts an sql statement as a configuration parameter where the values to be inserted are converters from patternlayout the message converter m is likely to always be included this allows attackers to manipulate the sql by entering crafted strings into input fields or headers of an application that are logged allowing unintended sql queries to be executed note this issue only affects x when specifically configured to use the jdbcappender which is not the default beginning in version the jdbcappender was re introduced with proper support for parameterized sql queries and further customization over the columns written to in logs apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails by design the jdbcappender in x accepts an sql statement as a configuration parameter where the values to be inserted are converters from patternlayout the message converter m is likely to always be included this allows attackers to manipulate the sql by entering crafted strings into input fields or headers of an application that are logged allowing unintended sql queries to be executed note this issue only affects x when specifically configured to use the jdbcappender which is not the default beginning in version the jdbcappender was re introduced with proper support for parameterized sql queries and further customization over the columns written to in logs apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions vulnerabilityurl
0
119,318
15,498,471,414
IssuesEvent
2021-03-11 06:31:03
jesus-collective/mobile
https://api.github.com/repos/jesus-collective/mobile
closed
Assignment input field overlapping other elements
Courses Design High Priority
The "write your assignment" text box appears overlaying other parts of the assignment page. This occurs in both the assignment and response lesson types. <img width="1436" alt="Screen Shot 2021-03-08 at 10 08 23 AM" src="https://user-images.githubusercontent.com/48295718/110356758-943c3700-8008-11eb-9bad-9fd871b2e078.png"> <img width="1440" alt="Screen Shot 2021-03-08 at 10 07 49 AM" src="https://user-images.githubusercontent.com/48295718/110356753-92727380-8008-11eb-865f-e7b4a1e831b1.png">
1.0
Assignment input field overlapping other elements - The "write your assignment" text box appears overlaying other parts of the assignment page. This occurs in both the assignment and response lesson types. <img width="1436" alt="Screen Shot 2021-03-08 at 10 08 23 AM" src="https://user-images.githubusercontent.com/48295718/110356758-943c3700-8008-11eb-9bad-9fd871b2e078.png"> <img width="1440" alt="Screen Shot 2021-03-08 at 10 07 49 AM" src="https://user-images.githubusercontent.com/48295718/110356753-92727380-8008-11eb-865f-e7b4a1e831b1.png">
non_test
assignment input field overlapping other elements the write your assignment text box appears overlaying other parts of the assignment page this occurs in both the assignment and response lesson types img width alt screen shot at am src img width alt screen shot at am src
0
191,165
6,826,482,006
IssuesEvent
2017-11-08 14:17:12
orange-alliance/the-orange-alliance
https://api.github.com/repos/orange-alliance/the-orange-alliance
opened
Add Advancement Tab to Event Page for Team Advancement
enhancement Medium Priority
Show Teams Advancing to the Next Event on the Advancement Tab, as well as link to that event
1.0
Add Advancement Tab to Event Page for Team Advancement - Show Teams Advancing to the Next Event on the Advancement Tab, as well as link to that event
non_test
add advancement tab to event page for team advancement show teams advancing to the next event on the advancement tab as well as link to that event
0
102,019
8,815,548,248
IssuesEvent
2018-12-29 20:05:21
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: weekly/tpcc-max failed
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/32734be7ac6413035c487f114deccabb5621e8d8 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=weekly/tpcc-max PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1072677&tab=buildLog ``` The test failed on master: test.go:703,cluster.go:1137,tpcc.go:110,cluster.go:1463,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1072677-weekly-tpcc-max:4 -- ./workload run tpcc --warehouses=1500 --histograms=logs/stats.json --ramp=5m0s --duration=144h0m0s {pgurl:1-3} returned: stderr: stdout: 0 15.0 29.3 3221.2 17179.9 25769.8 25769.8 delivery 1h55m5s 0 241.1 292.4 3221.2 6442.5 10200.5 14495.5 newOrder 1h55m5s 0 33.0 29.3 75.5 260.0 369.1 369.1 orderStatus 1h55m5s 0 229.1 293.0 1208.0 3355.4 6979.3 9126.8 payment 1h55m5s 0 26.0 29.3 671.1 1677.7 4160.7 4160.7 stockLevel 1h55m6s 0 27.0 29.3 10737.4 17179.9 18253.6 18253.6 delivery 1h55m6s 0 191.1 292.4 3489.7 6710.9 7784.6 10737.4 newOrder 1h55m6s 0 34.0 29.3 62.9 302.0 419.4 419.4 orderStatus 1h55m6s 0 219.1 293.0 1476.4 3221.2 5905.6 11274.3 payment 1h55m6s 0 25.0 29.3 771.8 2147.5 2281.7 2281.7 stockLevel Error: error in delivery: ERROR: no inbound stream connection (SQLSTATE XX000) Error: exit status 1 : exit status 1 test.go:703,cluster.go:1484,tpcc.go:120,tpcc.go:159: Goexit() was called ```
1.0
roachtest: weekly/tpcc-max failed - SHA: https://github.com/cockroachdb/cockroach/commits/32734be7ac6413035c487f114deccabb5621e8d8 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=weekly/tpcc-max PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1072677&tab=buildLog ``` The test failed on master: test.go:703,cluster.go:1137,tpcc.go:110,cluster.go:1463,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1072677-weekly-tpcc-max:4 -- ./workload run tpcc --warehouses=1500 --histograms=logs/stats.json --ramp=5m0s --duration=144h0m0s {pgurl:1-3} returned: stderr: stdout: 0 15.0 29.3 3221.2 17179.9 25769.8 25769.8 delivery 1h55m5s 0 241.1 292.4 3221.2 6442.5 10200.5 14495.5 newOrder 1h55m5s 0 33.0 29.3 75.5 260.0 369.1 369.1 orderStatus 1h55m5s 0 229.1 293.0 1208.0 3355.4 6979.3 9126.8 payment 1h55m5s 0 26.0 29.3 671.1 1677.7 4160.7 4160.7 stockLevel 1h55m6s 0 27.0 29.3 10737.4 17179.9 18253.6 18253.6 delivery 1h55m6s 0 191.1 292.4 3489.7 6710.9 7784.6 10737.4 newOrder 1h55m6s 0 34.0 29.3 62.9 302.0 419.4 419.4 orderStatus 1h55m6s 0 219.1 293.0 1476.4 3221.2 5905.6 11274.3 payment 1h55m6s 0 25.0 29.3 771.8 2147.5 2281.7 2281.7 stockLevel Error: error in delivery: ERROR: no inbound stream connection (SQLSTATE XX000) Error: exit status 1 : exit status 1 test.go:703,cluster.go:1484,tpcc.go:120,tpcc.go:159: Goexit() was called ```
test
roachtest weekly tpcc max failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests weekly tpcc max pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master test go cluster go tpcc go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity weekly tpcc max workload run tpcc warehouses histograms logs stats json ramp duration pgurl returned stderr stdout delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel error error in delivery error no inbound stream connection sqlstate error exit status exit status test go cluster go tpcc go tpcc go goexit was called
1
54,869
6,415,209,166
IssuesEvent
2017-08-08 12:15:18
LiskHQ/lisk-nano
https://api.github.com/repos/LiskHQ/lisk-nano
closed
Clean up unit test output
easy test
### Expected behaviour Unit test output should contain no warnings and errors ### Actual behaviour Currently, it contains this, even though we are not using PropTypes in our code: ``` Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' ```
1.0
Clean up unit test output - ### Expected behaviour Unit test output should contain no warnings and errors ### Actual behaviour Currently, it contains this, even though we are not using PropTypes in our code: ``` Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' Chrome 60.0.3112 (Mac OS X 10.12.6) WARN: 'Warning: Accessing PropTypes via the main React package is deprecated, and will be removed in React v16.0. Use the latest available v15.* prop-types package from npm instead. For info on usage, compatibility, migration and more, see https://fb.me/prop-types-docs' ```
test
clean up unit test output expected behaviour unit test output should contain no warnings and errors actual behaviour currently it contains this even though we are not using proptypes in our code chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see chrome mac os x warn warning accessing proptypes via the main react package is deprecated and will be removed in react use the latest available prop types package from npm instead for info on usage compatibility migration and more see
1
63,870
6,886,667,848
IssuesEvent
2017-11-21 20:19:47
arfc/publications
https://api.github.com/repos/arfc/publications
closed
Moltres init: In-depth scaling tests
Comp:Analysis Difficulty:3-Expert Priority:1-Critical Status:5-In Review Type:Test
> The proposed scaling test is very limited. Proper scaling tests would require weak and strong scaling tests, an assessment on the impact of the size of problem per core, a separate assessment of intra-node and extra-node scaling, etc. In addition, information should be provided on the employed numerical algorithms and on the parallelization strategy. The authors should either extend significantly the section, or remove it completely. Maybe other MOOSE results could presented, and then present some Moltres results in the same terms. This would show that the solvers and parallelization provided by MOOSE work well for the physics exhibited by MSRs.
1.0
Moltres init: In-depth scaling tests - > The proposed scaling test is very limited. Proper scaling tests would require weak and strong scaling tests, an assessment on the impact of the size of problem per core, a separate assessment of intra-node and extra-node scaling, etc. In addition, information should be provided on the employed numerical algorithms and on the parallelization strategy. The authors should either extend significantly the section, or remove it completely. Maybe other MOOSE results could presented, and then present some Moltres results in the same terms. This would show that the solvers and parallelization provided by MOOSE work well for the physics exhibited by MSRs.
test
moltres init in depth scaling tests the proposed scaling test is very limited proper scaling tests would require weak and strong scaling tests an assessment on the impact of the size of problem per core a separate assessment of intra node and extra node scaling etc in addition information should be provided on the employed numerical algorithms and on the parallelization strategy the authors should either extend significantly the section or remove it completely maybe other moose results could presented and then present some moltres results in the same terms this would show that the solvers and parallelization provided by moose work well for the physics exhibited by msrs
1
410,731
11,996,085,923
IssuesEvent
2020-04-08 16:10:31
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Upgrading 1.0.0 to 1.0.1 fails with "invalid config: apiVersion is not found."
area/kfctl kind/bug platform/onprem priority/p2
/kind bug **What steps did you take and what happened:** ➜ sudo kfctl build -f /Users/lalithvaka/installs/kfctl_upgrade_spec.yaml -V INFO[0000] Downloading /Users/lalithvaka/installs/kfctl_upgrade_spec.yaml to /tmp/532268644/update.yaml filename="v1alpha1/application_types.go:129" INFO[0000] Found KfCfg with matching name: kf-poc1 version: v1.0.0 at kf-poc1/kfctl_istio_dex.v1.0.0.yaml filename="kfupgrade/kfupgrade.go:195" INFO[0001] Downloading https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml to /tmp/408838771/tmp_app.yaml filename="loaders/loaders.go:71" Error: couldn't load KfUpgrade: (kubeflow.error): Code 500 with message: Encountered error while creating new KfApp kf-poc1: (kubeflow.error): Code 500 with message: Could not load https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml. Error: (kubeflow.error): Code 400 with message: invalid config: apiVersion is not found. Usage: kfctl build [flags] Flags: -f, --file string Static config file to use. Can be either a local path: export CONFIG=./kfctl_gcp_iap.yaml or a URL: export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_gcp_iap.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_istio_dex.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_aws.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_k8s_istio.v1.0.0.yaml kfctl build -V --file=${CONFIG} -h, --help help for build -V, --verbose verbose output default is false couldn't load KfUpgrade: (kubeflow.error): Code 500 with message: Encountered error while creating new KfApp kf-poc1: (kubeflow.error): Code 500 with message: Could not load https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml. Error: (kubeflow.error): Code 400 with message: invalid config: apiVersion is not found **What did you expect to happen:** **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] **Environment:** - Kubeflow version: 1.0.0 - kfctl version: (use `kfctl version`): kfctl v1.0.1-0-gf3edb9b - Kubernetes platform: (e.g. `minikube`) : onPrem - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`):
1.0
Upgrading 1.0.0 to 1.0.1 fails with "invalid config: apiVersion is not found." - /kind bug **What steps did you take and what happened:** ➜ sudo kfctl build -f /Users/lalithvaka/installs/kfctl_upgrade_spec.yaml -V INFO[0000] Downloading /Users/lalithvaka/installs/kfctl_upgrade_spec.yaml to /tmp/532268644/update.yaml filename="v1alpha1/application_types.go:129" INFO[0000] Found KfCfg with matching name: kf-poc1 version: v1.0.0 at kf-poc1/kfctl_istio_dex.v1.0.0.yaml filename="kfupgrade/kfupgrade.go:195" INFO[0001] Downloading https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml to /tmp/408838771/tmp_app.yaml filename="loaders/loaders.go:71" Error: couldn't load KfUpgrade: (kubeflow.error): Code 500 with message: Encountered error while creating new KfApp kf-poc1: (kubeflow.error): Code 500 with message: Could not load https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml. Error: (kubeflow.error): Code 400 with message: invalid config: apiVersion is not found. Usage: kfctl build [flags] Flags: -f, --file string Static config file to use. Can be either a local path: export CONFIG=./kfctl_gcp_iap.yaml or a URL: export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_gcp_iap.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_istio_dex.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_aws.v1.0.0.yaml export CONFIG=https://raw.githubusercontent.com/kubeflow/manifests/v1.0-branch/kfdef/kfctl_k8s_istio.v1.0.0.yaml kfctl build -V --file=${CONFIG} -h, --help help for build -V, --verbose verbose output default is false couldn't load KfUpgrade: (kubeflow.error): Code 500 with message: Encountered error while creating new KfApp kf-poc1: (kubeflow.error): Code 500 with message: Could not load https://github.com/kubeflow/manifests/blob/master/kfdef/kfctl_istio_dex.v1.0.1.yaml. Error: (kubeflow.error): Code 400 with message: invalid config: apiVersion is not found **What did you expect to happen:** **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] **Environment:** - Kubeflow version: 1.0.0 - kfctl version: (use `kfctl version`): kfctl v1.0.1-0-gf3edb9b - Kubernetes platform: (e.g. `minikube`) : onPrem - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`):
non_test
upgrading to fails with invalid config apiversion is not found kind bug what steps did you take and what happened ➜ sudo kfctl build f users lalithvaka installs kfctl upgrade spec yaml v info downloading users lalithvaka installs kfctl upgrade spec yaml to tmp update yaml filename application types go info found kfcfg with matching name kf version at kf kfctl istio dex yaml filename kfupgrade kfupgrade go info downloading to tmp tmp app yaml filename loaders loaders go error couldn t load kfupgrade kubeflow error code with message encountered error while creating new kfapp kf kubeflow error code with message could not load error kubeflow error code with message invalid config apiversion is not found usage kfctl build flags f file string static config file to use can be either a local path export config kfctl gcp iap yaml or a url export config export config export config export config kfctl build v file config h help help for build v verbose verbose output default is false couldn t load kfupgrade kubeflow error code with message encountered error while creating new kfapp kf kubeflow error code with message could not load error kubeflow error code with message invalid config apiversion is not found what did you expect to happen anything else you would like to add environment kubeflow version kfctl version use kfctl version kfctl kubernetes platform e g minikube onprem kubernetes version use kubectl version os e g from etc os release
0
19,013
3,737,502,554
IssuesEvent
2016-03-08 19:30:12
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
circleci: failed tests (14058): TestExplainTrace
test-failure
The following test appears to have failed: [#14058](https://circleci.com/gh/cockroachdb/cockroach/14058): ``` I160308 19:26:41.802651 storage/replica_command.go:453 auto-gc'ed f9258977 (1 intents) I160308 19:26:41.808979 storage/engine/rocksdb.go:141 closing in-memory rocksdb instance I160308 19:26:41.809087 http2_client.go:877 transport: http2Client.notifyError got notified that the client transport was broken EOF. I160308 19:26:42.809946 /go/src/google.golang.org/grpc/clientconn.go:463 grpc: Conn.resetTransport failed to create client transport: connection error: desc = "transport: dial tcp 127.0.0.1:45389: getsockopt: connection refused"; Reconnecting to "127.0.0.1:45389" I160308 19:26:42.810050 /go/src/google.golang.org/grpc/clientconn.go:501 grpc: Conn.transportMonitor exits due to: grpc: the client connection is closing --- FAIL: TestExplainTrace (1.51s) panic: runtime error: slice bounds out of range [recovered] panic: runtime error: slice bounds out of range [recovered] panic: runtime error: slice bounds out of range goroutine 26856 [running]: panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:464 +0x3ff testing.tRunner.func1(0xc82070f3b0) /usr/local/go/src/testing/testing.go:467 +0x22a panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:426 +0x521 github.com/cockroachdb/pq.(*conn).errRecover(0xc820793180, 0xc820fdba88) /go/src/github.com/cockroachdb/pq/error.go:482 +0x258 panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:426 +0x521 github.com/cockroachdb/pq.parseTs(0x0, 0xc820313fc0, 0x7, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:312 +0x14d7 github.com/cockroachdb/pq.textDecode(0xc8207933a8, 0xc8207931a6, 0x7, 0x1fa, 0xc80000045a, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:91 +0x63d github.com/cockroachdb/pq.decode(0xc8207933a8, 0xc8207931a6, 0x7, 0x1fa, 0x45a, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:62 +0xc0 github.com/cockroachdb/pq.(*rows).Next(0xc820016cf0, 0xc820016ea0, 0x9, 0x9, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/conn.go:1541 +0xbd0 database/sql.(*Rows).Next(0xc820ecc180, 0xc820313fb0) -- /go/src/github.com/cockroachdb/cockroach/sql/trace_test.go:98 +0x44f testing.tRunner(0xc82070f3b0, 0x28a9118) /usr/local/go/src/testing/testing.go:473 +0xdd created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0xae3 FAIL github.com/cockroachdb/cockroach/sql 105.354s ? github.com/cockroachdb/cockroach/sql/driver [no test files] === RUN TestEval --- PASS: TestEval (0.05s) === RUN TestEvalError --- PASS: TestEvalError (0.00s) === RUN TestEvalComparisonExprCaching --- PASS: TestEvalComparisonExprCaching (0.00s) === RUN TestSimilarEscape --- PASS: TestSimilarEscape (0.00s) === RUN TestQualifiedNameString ``` Please assign, take a look and update the issue accordingly.
1.0
circleci: failed tests (14058): TestExplainTrace - The following test appears to have failed: [#14058](https://circleci.com/gh/cockroachdb/cockroach/14058): ``` I160308 19:26:41.802651 storage/replica_command.go:453 auto-gc'ed f9258977 (1 intents) I160308 19:26:41.808979 storage/engine/rocksdb.go:141 closing in-memory rocksdb instance I160308 19:26:41.809087 http2_client.go:877 transport: http2Client.notifyError got notified that the client transport was broken EOF. I160308 19:26:42.809946 /go/src/google.golang.org/grpc/clientconn.go:463 grpc: Conn.resetTransport failed to create client transport: connection error: desc = "transport: dial tcp 127.0.0.1:45389: getsockopt: connection refused"; Reconnecting to "127.0.0.1:45389" I160308 19:26:42.810050 /go/src/google.golang.org/grpc/clientconn.go:501 grpc: Conn.transportMonitor exits due to: grpc: the client connection is closing --- FAIL: TestExplainTrace (1.51s) panic: runtime error: slice bounds out of range [recovered] panic: runtime error: slice bounds out of range [recovered] panic: runtime error: slice bounds out of range goroutine 26856 [running]: panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:464 +0x3ff testing.tRunner.func1(0xc82070f3b0) /usr/local/go/src/testing/testing.go:467 +0x22a panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:426 +0x521 github.com/cockroachdb/pq.(*conn).errRecover(0xc820793180, 0xc820fdba88) /go/src/github.com/cockroachdb/pq/error.go:482 +0x258 panic(0x1ab9aa0, 0xc8200100a0) /usr/local/go/src/runtime/panic.go:426 +0x521 github.com/cockroachdb/pq.parseTs(0x0, 0xc820313fc0, 0x7, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:312 +0x14d7 github.com/cockroachdb/pq.textDecode(0xc8207933a8, 0xc8207931a6, 0x7, 0x1fa, 0xc80000045a, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:91 +0x63d github.com/cockroachdb/pq.decode(0xc8207933a8, 0xc8207931a6, 0x7, 0x1fa, 0x45a, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/encode.go:62 +0xc0 github.com/cockroachdb/pq.(*rows).Next(0xc820016cf0, 0xc820016ea0, 0x9, 0x9, 0x0, 0x0) /go/src/github.com/cockroachdb/pq/conn.go:1541 +0xbd0 database/sql.(*Rows).Next(0xc820ecc180, 0xc820313fb0) -- /go/src/github.com/cockroachdb/cockroach/sql/trace_test.go:98 +0x44f testing.tRunner(0xc82070f3b0, 0x28a9118) /usr/local/go/src/testing/testing.go:473 +0xdd created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0xae3 FAIL github.com/cockroachdb/cockroach/sql 105.354s ? github.com/cockroachdb/cockroach/sql/driver [no test files] === RUN TestEval --- PASS: TestEval (0.05s) === RUN TestEvalError --- PASS: TestEvalError (0.00s) === RUN TestEvalComparisonExprCaching --- PASS: TestEvalComparisonExprCaching (0.00s) === RUN TestSimilarEscape --- PASS: TestSimilarEscape (0.00s) === RUN TestQualifiedNameString ``` Please assign, take a look and update the issue accordingly.
test
circleci failed tests testexplaintrace the following test appears to have failed storage replica command go auto gc ed intents storage engine rocksdb go closing in memory rocksdb instance client go transport notifyerror got notified that the client transport was broken eof go src google golang org grpc clientconn go grpc conn resettransport failed to create client transport connection error desc transport dial tcp getsockopt connection refused reconnecting to go src google golang org grpc clientconn go grpc conn transportmonitor exits due to grpc the client connection is closing fail testexplaintrace panic runtime error slice bounds out of range panic runtime error slice bounds out of range panic runtime error slice bounds out of range goroutine panic usr local go src runtime panic go testing trunner usr local go src testing testing go panic usr local go src runtime panic go github com cockroachdb pq conn errrecover go src github com cockroachdb pq error go panic usr local go src runtime panic go github com cockroachdb pq parsets go src github com cockroachdb pq encode go github com cockroachdb pq textdecode go src github com cockroachdb pq encode go github com cockroachdb pq decode go src github com cockroachdb pq encode go github com cockroachdb pq rows next go src github com cockroachdb pq conn go database sql rows next go src github com cockroachdb cockroach sql trace test go testing trunner usr local go src testing testing go created by testing runtests usr local go src testing testing go fail github com cockroachdb cockroach sql github com cockroachdb cockroach sql driver run testeval pass testeval run testevalerror pass testevalerror run testevalcomparisonexprcaching pass testevalcomparisonexprcaching run testsimilarescape pass testsimilarescape run testqualifiednamestring please assign take a look and update the issue accordingly
1
14,794
3,422,362,488
IssuesEvent
2015-12-08 22:41:16
metafizzy/isotope
https://api.github.com/repos/metafizzy/isotope
closed
How to check is isotope has been initialised?
test case required
I am adding new items on an ajax call, and I check each time I click to load new items if isotope has been initialised, the isotope layout happens laying all the new items as it should but the following never happens if($("#list").data('isotope')) { alert("hello"); } Is there anyway otherway to check if isotope has already been initialised?
1.0
How to check is isotope has been initialised? - I am adding new items on an ajax call, and I check each time I click to load new items if isotope has been initialised, the isotope layout happens laying all the new items as it should but the following never happens if($("#list").data('isotope')) { alert("hello"); } Is there anyway otherway to check if isotope has already been initialised?
test
how to check is isotope has been initialised i am adding new items on an ajax call and i check each time i click to load new items if isotope has been initialised the isotope layout happens laying all the new items as it should but the following never happens if list data isotope alert hello is there anyway otherway to check if isotope has already been initialised
1
275,615
23,925,387,035
IssuesEvent
2022-09-09 21:49:18
hashicorp/terraform-provider-google
https://api.github.com/repos/hashicorp/terraform-provider-google
opened
Failing test(s): TestAccDataprocMetastoreService_dataprocMetastoreServiceCmekTestExample
test failure
<!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since Aug 6 2022 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccDataprocMetastoreService_dataprocMetastoreServiceCmekTestExample <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloud_ProviderGoogleCloudGoogleProject/335482?buildTab=tests&expandedTest=-1290592027607407147 <!-- The error message that displays in the tests tab, for reference --> Message: ``` Error: Error waiting to create Service: Error waiting for Creating Service: Error code 13, message: an internal error has occurred with google_dataproc_metastore_service.default ```
1.0
Failing test(s): TestAccDataprocMetastoreService_dataprocMetastoreServiceCmekTestExample - <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since Aug 6 2022 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccDataprocMetastoreService_dataprocMetastoreServiceCmekTestExample <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloud_ProviderGoogleCloudGoogleProject/335482?buildTab=tests&expandedTest=-1290592027607407147 <!-- The error message that displays in the tests tab, for reference --> Message: ``` Error: Error waiting to create Service: Error waiting for Creating Service: Error code 13, message: an internal error has occurred with google_dataproc_metastore_service.default ```
test
failing test s testaccdataprocmetastoreservice dataprocmetastoreservicecmektestexample failure rate since aug impacted tests testaccdataprocmetastoreservice dataprocmetastoreservicecmektestexample nightly builds message error error waiting to create service error waiting for creating service error code message an internal error has occurred with google dataproc metastore service default
1
322,402
27,598,358,139
IssuesEvent
2023-03-09 08:21:14
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix jax_numpy_logic.test_jax_numpy_bitwise_and
JAX Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details>
1.0
Fix jax_numpy_logic.test_jax_numpy_bitwise_and - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4367837659/jobs/7639653664" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_bitwise_and[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-08T20:44:15.8716783Z E jax._src.traceback_util.UnfilteredStackTrace: TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8717170Z E 2023-03-08T20:44:15.8717488Z E The stack trace below excludes JAX-internal frames. 2023-03-08T20:44:15.8717851Z E The preceding is the original exception that occurred, unmodified. 2023-03-08T20:44:15.8718125Z E 2023-03-08T20:44:15.8718359Z E -------------------- 2023-03-08T20:44:15.8721853Z E TypeError: <lambda>() got some positional-only arguments passed as keyword arguments: 'x1, x2' 2023-03-08T20:44:15.8722239Z E Falsifying example: test_jax_numpy_bitwise_and( 2023-03-08T20:44:15.8730934Z E dtype_and_x=(['bool', 'bool'], [array(False), array(False)]), 2023-03-08T20:44:15.8731281Z E test_flags=FrontendFunctionTestFlags( 2023-03-08T20:44:15.8731565Z E num_positional_args=0, 2023-03-08T20:44:15.8731805Z E with_out=False, 2023-03-08T20:44:15.8732040Z E inplace=False, 2023-03-08T20:44:15.8732274Z E as_variable=[False], 2023-03-08T20:44:15.8732516Z E native_arrays=[False], 2023-03-08T20:44:15.8732737Z E ), 2023-03-08T20:44:15.8733103Z E fn_tree='ivy.functional.frontends.jax.numpy.bitwise_and', 2023-03-08T20:44:15.8733445Z E on_device='cpu', 2023-03-08T20:44:15.8733703Z E frontend='jax', 2023-03-08T20:44:15.8733903Z E ) 2023-03-08T20:44:15.8734077Z E 2023-03-08T20:44:15.8734560Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AAAAAAAAAAA=') as a decorator on your test case </details>
test
fix jax numpy logic test jax numpy bitwise and tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy bitwise and e jax src traceback util unfilteredstacktrace typeerror got some positional only arguments passed as keyword arguments e e the stack trace below excludes jax internal frames e the preceding is the original exception that occurred unmodified e e e typeerror got some positional only arguments passed as keyword arguments e falsifying example test jax numpy bitwise and e dtype and x e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e e fn tree ivy functional frontends jax numpy bitwise and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b aaaaaaaaaaa as a decorator on your test case failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy bitwise and e jax src traceback util unfilteredstacktrace typeerror got some positional only arguments passed as keyword arguments e e the stack trace below excludes jax internal frames e the preceding is the original exception that occurred unmodified e e e typeerror got some positional only arguments passed as keyword arguments e falsifying example test jax numpy bitwise and e dtype and x e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e e fn tree ivy functional frontends jax numpy bitwise and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b aaaaaaaaaaa as a decorator on your test case failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy bitwise and e jax src traceback util unfilteredstacktrace typeerror got some positional only arguments passed as keyword arguments e e the stack trace below excludes jax internal frames e the preceding is the original exception that occurred unmodified e e e typeerror got some positional only arguments passed as keyword arguments e falsifying example test jax numpy bitwise and e dtype and x e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e e fn tree ivy functional frontends jax numpy bitwise and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b aaaaaaaaaaa as a decorator on your test case
1
343,579
10,333,167,641
IssuesEvent
2019-09-03 03:59:36
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.skroutz.gr - Category menu doesn't work
browser-firefox-mobile engine-gecko priority-normal severity-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.skroutz.gr/s/6205892/Hyperdevotion-Noire-Goddess-Black-Heart-PSVita.html?from=account_favorites **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Category menu doesn't work **Steps to Reproduce**: I can't choose the category on top of the site by pressing on the product name, the scrypt doesn't work. on pc it works fine [![Screenshot Description](https://webcompat.com/uploads/2019/8/1991bae6-8e6d-4553-a08c-d4bc5543af5d-thumb.jpeg)](https://webcompat.com/uploads/2019/8/1991bae6-8e6d-4553-a08c-d4bc5543af5d.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190819150103</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.skroutz.gr - Category menu doesn't work - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.skroutz.gr/s/6205892/Hyperdevotion-Noire-Goddess-Black-Heart-PSVita.html?from=account_favorites **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Category menu doesn't work **Steps to Reproduce**: I can't choose the category on top of the site by pressing on the product name, the scrypt doesn't work. on pc it works fine [![Screenshot Description](https://webcompat.com/uploads/2019/8/1991bae6-8e6d-4553-a08c-d4bc5543af5d-thumb.jpeg)](https://webcompat.com/uploads/2019/8/1991bae6-8e6d-4553-a08c-d4bc5543af5d.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190819150103</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
category menu doesn t work url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description category menu doesn t work steps to reproduce i can t choose the category on top of the site by pressing on the product name the scrypt doesn t work on pc it works fine browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️
0
133,572
12,544,682,378
IssuesEvent
2020-06-05 17:37:15
Azure-Samples/Cognitive-Services-Voice-Assistant
https://api.github.com/repos/Azure-Samples/Cognitive-Services-Voice-Assistant
closed
Add documentation around commands
Devices Console Client (C++) documentation
### This issue is for a: (mark with an `x`) ``` - [ ] bug report -> please search issues before submitting - [ ] feature request - [ x] documentation issue or request - [ ] regression (a behavior that used to work and stopped in a new release) ``` There is default behavior that starts KWS so we should document that. Also there are commands that print out if you have a console interface which show examples of how to use the basic API commands
1.0
Add documentation around commands - ### This issue is for a: (mark with an `x`) ``` - [ ] bug report -> please search issues before submitting - [ ] feature request - [ x] documentation issue or request - [ ] regression (a behavior that used to work and stopped in a new release) ``` There is default behavior that starts KWS so we should document that. Also there are commands that print out if you have a console interface which show examples of how to use the basic API commands
non_test
add documentation around commands this issue is for a mark with an x bug report please search issues before submitting feature request documentation issue or request regression a behavior that used to work and stopped in a new release there is default behavior that starts kws so we should document that also there are commands that print out if you have a console interface which show examples of how to use the basic api commands
0
771,967
27,100,013,535
IssuesEvent
2023-02-15 07:51:37
therealbluepandabear/PixaPencil_Classic
https://api.github.com/repos/therealbluepandabear/PixaPencil_Classic
closed
[B] Zooming out sometimes resets the grid
🐛 bug mid priority difficulty: hard
Zooming out sometimes resets the grid, I think it's a rounding error.
1.0
[B] Zooming out sometimes resets the grid - Zooming out sometimes resets the grid, I think it's a rounding error.
non_test
zooming out sometimes resets the grid zooming out sometimes resets the grid i think it s a rounding error
0
172,160
13,264,101,179
IssuesEvent
2020-08-21 02:36:59
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
An empty folder displays when searching one existing blob in one folder under 'Active blobs and blobs without current' view
:gear: blobs 🧪 testing
**Storage Explorer Version:** 1.15.0-dev **Build**: 20200820.5 **Branch**: master **Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one Non-ADLS Gen2 storage account (make sure the blob versioning is enabled) -> Blob Containers. 2. Create one blob container -> Upload one folder with blobs to it. 3. Switch to 'Active blobs and blobs without current' view. 4. Open the folder -> Type one blob's name in the 'Search by prefix(case-sensitive)' box. 5. Check the result. **Expect Experience:** The bob shows well in the editor. **Actual Experience:** An empty folder shows in the editor. ![image](https://user-images.githubusercontent.com/54055206/90749181-b76df600-e305-11ea-9e63-0fff055f9869.png) **More info:** This issue also reproduces under 'All blobs and blobs without current' view.
1.0
An empty folder displays when searching one existing blob in one folder under 'Active blobs and blobs without current' view - **Storage Explorer Version:** 1.15.0-dev **Build**: 20200820.5 **Branch**: master **Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one Non-ADLS Gen2 storage account (make sure the blob versioning is enabled) -> Blob Containers. 2. Create one blob container -> Upload one folder with blobs to it. 3. Switch to 'Active blobs and blobs without current' view. 4. Open the folder -> Type one blob's name in the 'Search by prefix(case-sensitive)' box. 5. Check the result. **Expect Experience:** The bob shows well in the editor. **Actual Experience:** An empty folder shows in the editor. ![image](https://user-images.githubusercontent.com/54055206/90749181-b76df600-e305-11ea-9e63-0fff055f9869.png) **More info:** This issue also reproduces under 'All blobs and blobs without current' view.
test
an empty folder displays when searching one existing blob in one folder under active blobs and blobs without current view storage explorer version dev build branch master platform os windows centos core macos catalina architecture regression from not a regression steps to reproduce expand one non adls storage account make sure the blob versioning is enabled blob containers create one blob container upload one folder with blobs to it switch to active blobs and blobs without current view open the folder type one blob s name in the search by prefix case sensitive box check the result expect experience the bob shows well in the editor actual experience an empty folder shows in the editor more info this issue also reproduces under all blobs and blobs without current view
1
19,648
14,369,018,439
IssuesEvent
2020-12-01 09:13:15
cpn-io/cpn-js
https://api.github.com/repos/cpn-io/cpn-js
closed
Front-end/Simulation: show progress counter in replication simulation
fixed usability issue
Replication simulation can take a long time (several minutes, even up to an hour). Progress of replication can be tracked as follows: each replication run creates a subdirectory sim_X Show and update progress <current> / <total sims> on the screen "Replication processing..." Also show message "This may take several minutes..."
True
Front-end/Simulation: show progress counter in replication simulation - Replication simulation can take a long time (several minutes, even up to an hour). Progress of replication can be tracked as follows: each replication run creates a subdirectory sim_X Show and update progress <current> / <total sims> on the screen "Replication processing..." Also show message "This may take several minutes..."
non_test
front end simulation show progress counter in replication simulation replication simulation can take a long time several minutes even up to an hour progress of replication can be tracked as follows each replication run creates a subdirectory sim x show and update progress on the screen replication processing also show message this may take several minutes
0
201,150
15,801,967,336
IssuesEvent
2021-04-03 07:22:37
samleewy/ped
https://api.github.com/repos/samleewy/ped
opened
"uncompleted" command not in UG
severity.Low type.DocumentationBug
I found out that typing `uncompleted` switches me to the uncompleted tab. This was not included in the UG under "Switching between tabs". ![image.png](https://raw.githubusercontent.com/samleewy/ped/main/files/2cf8f4cd-299a-43c6-9f18-8871d9de00a6.png) <!--session: 1617429558329-c98042ac-2dd2-4929-9aa4-0d4bc108a360-->
1.0
"uncompleted" command not in UG - I found out that typing `uncompleted` switches me to the uncompleted tab. This was not included in the UG under "Switching between tabs". ![image.png](https://raw.githubusercontent.com/samleewy/ped/main/files/2cf8f4cd-299a-43c6-9f18-8871d9de00a6.png) <!--session: 1617429558329-c98042ac-2dd2-4929-9aa4-0d4bc108a360-->
non_test
uncompleted command not in ug i found out that typing uncompleted switches me to the uncompleted tab this was not included in the ug under switching between tabs
0
148,295
11,846,450,899
IssuesEvent
2020-03-24 10:13:10
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Views: Support contributable welcome view content
api api-finalization feature-request on-testplan workbench-views
### Problem The need to have meaningful and actionable content rendered on empty views has been arising across the workbench: - SCM: https://github.com/microsoft/vscode/issues/61491 - Debug: https://github.com/microsoft/vscode/issues/85548 - Custom Views: https://github.com/microsoft/vscode-remote-release/issues/1349 ### Proposal Let's enable the following in the core: 1. Add an overridable `isEmpty(): boolean` method to `ViewPane` which would let view panes dictate when they are empty, as well as `onDidChangeEmpty` event 2. Create a core contribution point for empty view content, supporting very simple Markdown (links only) 3. Enhance the simple Markdown to be able to detect isolated command links and render them as buttons 4. Render all empty view contributions in a `ViewPane` when `isEmpty() === true` We can then update our API by exposing an extension contribution point for the same core contribution point created above. Here's a couple of proposals: **contributes.documentation.VIEWID** ```json "contributes": { "documentation": { "VIEWID": [{ "label": "MARKDOWN", "when": "CONDITION" }] } } ``` Alternatives for `documentation`: `viewhelp`, `emptyview`. Alternatives for `label`: `text`, `markdown`, `description`. **contributes.documentation.view** ```json "contributes": { "documentation": { "view": [{ "label": "MARKDOWN", "when": "VIEWID && CONDITION" }] } } ``` --- cc @sbatten for views cc @isidorn for debug cc @sandy081, @alexr00 for custom views cc @bowdenk7 for scm cc @jrieken for input on API conventions
1.0
Views: Support contributable welcome view content - ### Problem The need to have meaningful and actionable content rendered on empty views has been arising across the workbench: - SCM: https://github.com/microsoft/vscode/issues/61491 - Debug: https://github.com/microsoft/vscode/issues/85548 - Custom Views: https://github.com/microsoft/vscode-remote-release/issues/1349 ### Proposal Let's enable the following in the core: 1. Add an overridable `isEmpty(): boolean` method to `ViewPane` which would let view panes dictate when they are empty, as well as `onDidChangeEmpty` event 2. Create a core contribution point for empty view content, supporting very simple Markdown (links only) 3. Enhance the simple Markdown to be able to detect isolated command links and render them as buttons 4. Render all empty view contributions in a `ViewPane` when `isEmpty() === true` We can then update our API by exposing an extension contribution point for the same core contribution point created above. Here's a couple of proposals: **contributes.documentation.VIEWID** ```json "contributes": { "documentation": { "VIEWID": [{ "label": "MARKDOWN", "when": "CONDITION" }] } } ``` Alternatives for `documentation`: `viewhelp`, `emptyview`. Alternatives for `label`: `text`, `markdown`, `description`. **contributes.documentation.view** ```json "contributes": { "documentation": { "view": [{ "label": "MARKDOWN", "when": "VIEWID && CONDITION" }] } } ``` --- cc @sbatten for views cc @isidorn for debug cc @sandy081, @alexr00 for custom views cc @bowdenk7 for scm cc @jrieken for input on API conventions
test
views support contributable welcome view content problem the need to have meaningful and actionable content rendered on empty views has been arising across the workbench scm debug custom views proposal let s enable the following in the core add an overridable isempty boolean method to viewpane which would let view panes dictate when they are empty as well as ondidchangeempty event create a core contribution point for empty view content supporting very simple markdown links only enhance the simple markdown to be able to detect isolated command links and render them as buttons render all empty view contributions in a viewpane when isempty true we can then update our api by exposing an extension contribution point for the same core contribution point created above here s a couple of proposals contributes documentation viewid json contributes documentation viewid label markdown when condition alternatives for documentation viewhelp emptyview alternatives for label text markdown description contributes documentation view json contributes documentation view label markdown when viewid condition cc sbatten for views cc isidorn for debug cc for custom views cc for scm cc jrieken for input on api conventions
1
444,813
31,149,165,908
IssuesEvent
2023-08-16 08:46:07
openearth/aeolis-python
https://api.github.com/repos/openearth/aeolis-python
closed
Readthedocs does not display model.py in the overview of model code.
documentation
https://aeolis.readthedocs.io/en/latest/_modules/index.html I am missing a link to 'model'. Maybe other modules are not displayed neither.
1.0
Readthedocs does not display model.py in the overview of model code. - https://aeolis.readthedocs.io/en/latest/_modules/index.html I am missing a link to 'model'. Maybe other modules are not displayed neither.
non_test
readthedocs does not display model py in the overview of model code i am missing a link to model maybe other modules are not displayed neither
0
304,065
26,250,262,184
IssuesEvent
2023-01-05 18:37:22
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
opened
New Testimonial for Eddie Jaoude
testimonial
### Name eddiejaoude ### Title Great Man ### Description **Eddiejaoude is an amazing person who teaches developers how to contribute to open-source projects, and how to start Open Source.**
1.0
New Testimonial for Eddie Jaoude - ### Name eddiejaoude ### Title Great Man ### Description **Eddiejaoude is an amazing person who teaches developers how to contribute to open-source projects, and how to start Open Source.**
test
new testimonial for eddie jaoude name eddiejaoude title great man description eddiejaoude is an amazing person who teaches developers how to contribute to open source projects and how to start open source
1
184,658
6,715,059,155
IssuesEvent
2017-10-13 19:27:54
mlibrary/search
https://api.github.com/repos/mlibrary/search
closed
Performance is poor.
Priority: High Status: Maintenance Type: Performance
An audit by Chrome rates performance 16/100. ![capture](https://user-images.githubusercontent.com/1678665/30748219-9c6a3a54-9f7d-11e7-9d50-af4f7d66d660.PNG) **Tasks** - [x] Add filters as a batch - [ ] Add records as a batch - [ ] Add fields as a batch - [ ] Look into other areas for performance improvements - [ ] compress (gzip) spectrum requests once we are in staging/production environment.
1.0
Performance is poor. - An audit by Chrome rates performance 16/100. ![capture](https://user-images.githubusercontent.com/1678665/30748219-9c6a3a54-9f7d-11e7-9d50-af4f7d66d660.PNG) **Tasks** - [x] Add filters as a batch - [ ] Add records as a batch - [ ] Add fields as a batch - [ ] Look into other areas for performance improvements - [ ] compress (gzip) spectrum requests once we are in staging/production environment.
non_test
performance is poor an audit by chrome rates performance tasks add filters as a batch add records as a batch add fields as a batch look into other areas for performance improvements compress gzip spectrum requests once we are in staging production environment
0
80,796
15,589,001,442
IssuesEvent
2021-03-18 07:23:03
soumya132/pomscan
https://api.github.com/repos/soumya132/pomscan
closed
CVE-2019-11272 (High) detected in spring-security-core-4.1.1.RELEASE.jar
security vulnerability
## CVE-2019-11272 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-core-4.1.1.RELEASE.jar</b></p></summary> <p>spring-security-core</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: pomscan/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-core/4.1.1.RELEASE/spring-security-core-4.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library) - spring-security-config-4.1.1.RELEASE.jar - :x: **spring-security-core-4.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Security, versions 4.2.x up to 4.2.12, and older unsupported versions support plain text passwords using PlaintextPasswordEncoder. If an application using an affected version of Spring Security is leveraging PlaintextPasswordEncoder and a user has a null encoded password, a malicious user (or attacker) can authenticate using a password of "null". <p>Publish Date: 2019-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11272>CVE-2019-11272</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11272">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11272</a></p> <p>Release Date: 2019-06-26</p> <p>Fix Resolution: org.springframework.security:spring-security-core:4.2.13.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-11272 (High) detected in spring-security-core-4.1.1.RELEASE.jar - ## CVE-2019-11272 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-core-4.1.1.RELEASE.jar</b></p></summary> <p>spring-security-core</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: pomscan/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-core/4.1.1.RELEASE/spring-security-core-4.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library) - spring-security-config-4.1.1.RELEASE.jar - :x: **spring-security-core-4.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Security, versions 4.2.x up to 4.2.12, and older unsupported versions support plain text passwords using PlaintextPasswordEncoder. If an application using an affected version of Spring Security is leveraging PlaintextPasswordEncoder and a user has a null encoded password, a malicious user (or attacker) can authenticate using a password of "null". <p>Publish Date: 2019-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11272>CVE-2019-11272</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11272">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11272</a></p> <p>Release Date: 2019-06-26</p> <p>Fix Resolution: org.springframework.security:spring-security-core:4.2.13.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in spring security core release jar cve high severity vulnerability vulnerable library spring security core release jar spring security core library home page a href path to dependency file pomscan pom xml path to vulnerable library home wss scanner repository org springframework security spring security core release spring security core release jar dependency hierarchy spring boot starter security release jar root library spring security config release jar x spring security core release jar vulnerable library found in base branch master vulnerability details spring security versions x up to and older unsupported versions support plain text passwords using plaintextpasswordencoder if an application using an affected version of spring security is leveraging plaintextpasswordencoder and a user has a null encoded password a malicious user or attacker can authenticate using a password of null publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security core release step up your open source security game with whitesource
0
353,083
10,548,679,602
IssuesEvent
2019-10-03 06:41:35
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.halfords.com - site is not usable
browser-firefox-tablet engine-gecko priority-normal type-tracking-protection-basic
<!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 5.1.1; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.halfords.com/webapp/wcs/stores/servlet/SearchRouter?storeId=10001&catalogId=10151&tabNo=1&action=listrefine&pageNo=1&pageSize=42&sort=we_recommend&srch=car+seat+covers **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 5.1.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: filter not refreshing results when sorting by price, rating etc. **Steps to Reproduce**: Searched for car seat covers. [![Screenshot Description](https://webcompat.com/uploads/2019/9/6e601f3e-fcd8-41c9-b60e-abb3b2ffbbba-thumb.jpeg)](https://webcompat.com/uploads/2019/9/6e601f3e-fcd8-41c9-b60e-abb3b2ffbbba.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190923132102</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (basic)</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "The resource at https://libs.coremetrics.com/eluminate.js was blocked because content blocking is enabled." {file: "https://www.halfords.com/webapp/wcs/stores/servlet/SearchRouter?storeId=10001&catalogId=10151&tabNo=1&action=listrefine&pageNo=1&pageSize=42&sort=we_recommend&srch=car+seat+covers" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.halfords.com - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 5.1.1; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.halfords.com/webapp/wcs/stores/servlet/SearchRouter?storeId=10001&catalogId=10151&tabNo=1&action=listrefine&pageNo=1&pageSize=42&sort=we_recommend&srch=car+seat+covers **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 5.1.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: filter not refreshing results when sorting by price, rating etc. **Steps to Reproduce**: Searched for car seat covers. [![Screenshot Description](https://webcompat.com/uploads/2019/9/6e601f3e-fcd8-41c9-b60e-abb3b2ffbbba-thumb.jpeg)](https://webcompat.com/uploads/2019/9/6e601f3e-fcd8-41c9-b60e-abb3b2ffbbba.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190923132102</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (basic)</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "The resource at https://libs.coremetrics.com/eluminate.js was blocked because content blocking is enabled." {file: "https://www.halfords.com/webapp/wcs/stores/servlet/SearchRouter?storeId=10001&catalogId=10151&tabNo=1&action=listrefine&pageNo=1&pageSize=42&sort=we_recommend&srch=car+seat+covers" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
site is not usable url browser version firefox mobile tablet operating system android tested another browser no problem type site is not usable description filter not refreshing results when sorting by price rating etc steps to reproduce searched for car seat covers browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked true basic console messages from with ❤️
0
277,562
21,048,799,061
IssuesEvent
2022-03-31 18:40:31
akebu6/JetBrains-Academy-Projects
https://api.github.com/repos/akebu6/JetBrains-Academy-Projects
closed
add parking lot project to readme
documentation good first issue
include the parking lot project to readme and add its description
1.0
add parking lot project to readme - include the parking lot project to readme and add its description
non_test
add parking lot project to readme include the parking lot project to readme and add its description
0
211,721
16,357,348,549
IssuesEvent
2021-05-14 01:46:35
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
fix TestFunctional/parallel/LoadImage for docker driver with containerd runtime
co/runtime/containerd kind/failing-test
TestFunctional/parallel/LoadImage is failing for docker/containerd Example of logs: https://storage.googleapis.com/minikube-builds/logs/master/c31bd57/Docker_Linux_containerd.html#fail_TestFunctional%2fparallel%2fLoadImage Full log: https://storage.googleapis.com/minikube-builds/logs/master/c31bd57/Docker_Linux_containerdout.txt `minikube image load <image>` returns ok, but then `minikue ssh -- sudo crictl inspecti <image>` can not find the image.
1.0
fix TestFunctional/parallel/LoadImage for docker driver with containerd runtime - TestFunctional/parallel/LoadImage is failing for docker/containerd Example of logs: https://storage.googleapis.com/minikube-builds/logs/master/c31bd57/Docker_Linux_containerd.html#fail_TestFunctional%2fparallel%2fLoadImage Full log: https://storage.googleapis.com/minikube-builds/logs/master/c31bd57/Docker_Linux_containerdout.txt `minikube image load <image>` returns ok, but then `minikue ssh -- sudo crictl inspecti <image>` can not find the image.
test
fix testfunctional parallel loadimage for docker driver with containerd runtime testfunctional parallel loadimage is failing for docker containerd example of logs full log minikube image load returns ok but then minikue ssh sudo crictl inspecti can not find the image
1
200,260
22,739,536,988
IssuesEvent
2022-07-07 01:24:03
matomo-org/matomo
https://api.github.com/repos/matomo-org/matomo
closed
Use SSL to download plugins from the Marketplace
Task c: Security
The goal of this issue is to have Piwik core download plugins and themes from the Marketplace over SSL when supported. If upgrade of plugins fail over HTTPS, over fallback to HTTP update of plugins (helping users keeping plugins up to date is vital) Follows up: #6441, Refs #7488
True
Use SSL to download plugins from the Marketplace - The goal of this issue is to have Piwik core download plugins and themes from the Marketplace over SSL when supported. If upgrade of plugins fail over HTTPS, over fallback to HTTP update of plugins (helping users keeping plugins up to date is vital) Follows up: #6441, Refs #7488
non_test
use ssl to download plugins from the marketplace the goal of this issue is to have piwik core download plugins and themes from the marketplace over ssl when supported if upgrade of plugins fail over https over fallback to http update of plugins helping users keeping plugins up to date is vital follows up refs
0
13,627
23,444,621,673
IssuesEvent
2022-08-15 18:17:35
NASA-PDS/registry-mgr
https://api.github.com/repos/NASA-PDS/registry-mgr
opened
As a user I want to delete whole bundles at once
requirement needs:triage
<!-- For more information on how to populate this new feature request, see the PDS Wiki on User Story Development: https://github.com/NASA-PDS/nasa-pds.github.io/wiki/Issue-Tracking#user-story-development --> ## 💪 Motivation ...so that I can avoid being forced to run a delete command for every lidvid in a bundle ## 📖 Additional Details If one harvests an initial bundle by directory that includes several other bundles/products, it becomes impossible to efficiently delete products using packageId. In this case, it would be nice if the delete-data command allowed a flag that followed the same logic as set-archive-status, which is that it follows references all the way down to the product level if the provided lidvid is for a bundle or collection. ## ⚖️ Acceptance Criteria If I run `registry-manager delete-data -lidvid urn:nasa:pds:somelidvid` where that lidvid represents a bundle or collection, then registry manager will perform a delete for every primary product within that bundle or collection (including refs). <!-- For Internal Dev Team Use --> ## ⚙️ Engineering Details <!-- Provide some design / implementation details and/or a sub-task checklist as needed. Convert issue to Epic if estimate is outside the scope of 1 sprint. -->
1.0
As a user I want to delete whole bundles at once - <!-- For more information on how to populate this new feature request, see the PDS Wiki on User Story Development: https://github.com/NASA-PDS/nasa-pds.github.io/wiki/Issue-Tracking#user-story-development --> ## 💪 Motivation ...so that I can avoid being forced to run a delete command for every lidvid in a bundle ## 📖 Additional Details If one harvests an initial bundle by directory that includes several other bundles/products, it becomes impossible to efficiently delete products using packageId. In this case, it would be nice if the delete-data command allowed a flag that followed the same logic as set-archive-status, which is that it follows references all the way down to the product level if the provided lidvid is for a bundle or collection. ## ⚖️ Acceptance Criteria If I run `registry-manager delete-data -lidvid urn:nasa:pds:somelidvid` where that lidvid represents a bundle or collection, then registry manager will perform a delete for every primary product within that bundle or collection (including refs). <!-- For Internal Dev Team Use --> ## ⚙️ Engineering Details <!-- Provide some design / implementation details and/or a sub-task checklist as needed. Convert issue to Epic if estimate is outside the scope of 1 sprint. -->
non_test
as a user i want to delete whole bundles at once for more information on how to populate this new feature request see the pds wiki on user story development 💪 motivation so that i can avoid being forced to run a delete command for every lidvid in a bundle 📖 additional details if one harvests an initial bundle by directory that includes several other bundles products it becomes impossible to efficiently delete products using packageid in this case it would be nice if the delete data command allowed a flag that followed the same logic as set archive status which is that it follows references all the way down to the product level if the provided lidvid is for a bundle or collection ⚖️ acceptance criteria if i run registry manager delete data lidvid urn nasa pds somelidvid where that lidvid represents a bundle or collection then registry manager will perform a delete for every primary product within that bundle or collection including refs ⚙️ engineering details provide some design implementation details and or a sub task checklist as needed convert issue to epic if estimate is outside the scope of sprint
0
275,034
23,890,384,177
IssuesEvent
2022-09-08 10:59:45
stores-cedcommerce/HSL-Home-page-design
https://api.github.com/repos/stores-cedcommerce/HSL-Home-page-design
closed
The policy text needed to be in capitalize form.
Footer section Desktop Content Type / typo Ready to test fixed
**Actual result:** The text policy is coming in lowercase. ![image](https://user-images.githubusercontent.com/102131636/188592784-92623c04-beaf-4bd8-9dc4-634da31bd896.png) **Expected result:** The policy text needed to be in capitalize form.
1.0
The policy text needed to be in capitalize form. - **Actual result:** The text policy is coming in lowercase. ![image](https://user-images.githubusercontent.com/102131636/188592784-92623c04-beaf-4bd8-9dc4-634da31bd896.png) **Expected result:** The policy text needed to be in capitalize form.
test
the policy text needed to be in capitalize form actual result the text policy is coming in lowercase expected result the policy text needed to be in capitalize form
1
193,543
14,656,269,470
IssuesEvent
2020-12-28 13:02:28
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
hashicorp/vault-plugin-secrets-azure: backend_test.go; 3 LoC
fresh test tiny
Found a possible issue in [hashicorp/vault-plugin-secrets-azure](https://www.github.com/hashicorp/vault-plugin-secrets-azure) at [backend_test.go](https://github.com/hashicorp/vault-plugin-secrets-azure/blob/1629d98b3b9079f7ec895d378445136b06033f88/backend_test.go#L213-L215) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to keyID was used in a composite literal at line 214 [Click here to see the code in its original context.](https://github.com/hashicorp/vault-plugin-secrets-azure/blob/1629d98b3b9079f7ec895d378445136b06033f88/backend_test.go#L213-L215) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for keyID := range m.passwords { creds = append(creds, graphrbac.PasswordCredential{KeyID: &keyID}) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 1629d98b3b9079f7ec895d378445136b06033f88
1.0
hashicorp/vault-plugin-secrets-azure: backend_test.go; 3 LoC - Found a possible issue in [hashicorp/vault-plugin-secrets-azure](https://www.github.com/hashicorp/vault-plugin-secrets-azure) at [backend_test.go](https://github.com/hashicorp/vault-plugin-secrets-azure/blob/1629d98b3b9079f7ec895d378445136b06033f88/backend_test.go#L213-L215) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to keyID was used in a composite literal at line 214 [Click here to see the code in its original context.](https://github.com/hashicorp/vault-plugin-secrets-azure/blob/1629d98b3b9079f7ec895d378445136b06033f88/backend_test.go#L213-L215) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for keyID := range m.passwords { creds = append(creds, graphrbac.PasswordCredential{KeyID: &keyID}) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 1629d98b3b9079f7ec895d378445136b06033f88
test
hashicorp vault plugin secrets azure backend test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to keyid was used in a composite literal at line click here to show the line s of go which triggered the analyzer go for keyid range m passwords creds append creds graphrbac passwordcredential keyid keyid leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
98,442
20,734,327,584
IssuesEvent
2022-03-14 12:23:52
jvegax/Acme-Toolkits
https://api.github.com/repos/jvegax/Acme-Toolkits
opened
Task-046 : Operations by inventors on toolkits
code 🧑‍💻
- List their own toolkits. - Show their own toolkits, including their prices, their components, and their tools.
1.0
Task-046 : Operations by inventors on toolkits - - List their own toolkits. - Show their own toolkits, including their prices, their components, and their tools.
non_test
task operations by inventors on toolkits list their own toolkits show their own toolkits including their prices their components and their tools
0
68,035
7,086,147,755
IssuesEvent
2018-01-11 13:40:16
apiaryio/dredd
https://api.github.com/repos/apiaryio/dredd
closed
Tests for managing child processes fail on Travis CI Trusty
bug flaky test
The tests for managing child processes ([file](https://github.com/apiaryio/dredd/blob/master/test/integration/child-process-test.coffee)) are extremely flaky or almost consistently failing on Travis CI if the build is on `trusty` infrastructure ([docs](https://docs.travis-ci.com/user/reference/trusty#Container-based-with-sudo%3A-false), Docker-based). The move to the `trusty` infrastructure was introduced in the #775, because without being on `trusty` it's not possible to get Python 3 `pip` working on non-Python Travis CI build. To hotfix the situation, https://github.com/apiaryio/dredd/pull/829 was created to remove Python from the build, but this isn't a solution. Dredd's \*NIX tests should run correctly on any \*NIX-based infrastructure, which ensures Dredd is stable on any standard environment. 61cdfa5 needs to be reverted, otherwise consistency of docs isn't ensured in the CI. `trusty` is Docker-based, which means not being able to run the tests on `trusty` could be a blocker for #422, #784, and possibly other issues.
1.0
Tests for managing child processes fail on Travis CI Trusty - The tests for managing child processes ([file](https://github.com/apiaryio/dredd/blob/master/test/integration/child-process-test.coffee)) are extremely flaky or almost consistently failing on Travis CI if the build is on `trusty` infrastructure ([docs](https://docs.travis-ci.com/user/reference/trusty#Container-based-with-sudo%3A-false), Docker-based). The move to the `trusty` infrastructure was introduced in the #775, because without being on `trusty` it's not possible to get Python 3 `pip` working on non-Python Travis CI build. To hotfix the situation, https://github.com/apiaryio/dredd/pull/829 was created to remove Python from the build, but this isn't a solution. Dredd's \*NIX tests should run correctly on any \*NIX-based infrastructure, which ensures Dredd is stable on any standard environment. 61cdfa5 needs to be reverted, otherwise consistency of docs isn't ensured in the CI. `trusty` is Docker-based, which means not being able to run the tests on `trusty` could be a blocker for #422, #784, and possibly other issues.
test
tests for managing child processes fail on travis ci trusty the tests for managing child processes are extremely flaky or almost consistently failing on travis ci if the build is on trusty infrastructure docker based the move to the trusty infrastructure was introduced in the because without being on trusty it s not possible to get python pip working on non python travis ci build to hotfix the situation was created to remove python from the build but this isn t a solution dredd s nix tests should run correctly on any nix based infrastructure which ensures dredd is stable on any standard environment needs to be reverted otherwise consistency of docs isn t ensured in the ci trusty is docker based which means not being able to run the tests on trusty could be a blocker for and possibly other issues
1
281,861
24,426,808,595
IssuesEvent
2022-10-06 03:59:17
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: unoptimized-query-oracle/disable-rules=all failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-queries
roachtest.unoptimized-query-oracle/disable-rules=all [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6775433?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6775433?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=all) on master @ [0eaeeb773474716753781289788fdd087fb9b166](https://github.com/cockroachdb/cockroach/commits/0eaeeb773474716753781289788fdd087fb9b166): ``` |    }, ""), |    strings.Join({ |    ... // 769 identical bytes |    "0C18A117E67086FEEC1EABF114477C1E4C15123643403EFF3C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |    strings.Join({ |    ... // 953 identical bytes |    "5C150B766818ABAF0C120518B2F2EAEFE414CD73A1320CEE8C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |    strings.Join({ |    ... // 2965 identical bytes |    "3C1E498DCAF61A4E141C0587B873586F1C1C019D78E4F37D2C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |   } | sql: SELECT | '1983-07-24 03:35:34.000493':::TIMESTAMP AS col_4000, | tab_1393.col3_8 AS col_4001, | (-1.0717353820800781):::FLOAT8 AS col_4002, | 0:::OID AS col_4003, | tab_1394.col3_13 AS col_4004, | NULL AS col_4005, | tab_1393.col3_2 AS col_4006, | tab_1394.col3_1 AS col_4007, | (-7318860970.773327295):::DECIMAL AS col_4008, | tab_1393.col3_11 AS col_4009, | tab_1393.col3_7 AS col_4010, | tab_1394.col3_0 AS col_4011, | tab_1394.col3_10 AS col_4012, | tab_1393.col3_5 AS col_4013, | tab_1393.col3_0 AS col_4014, | tab_1394.col3_14 AS col_4015, | '1987-12-04 01:57:20.000745':::TIMESTAMP AS col_4016, | '!':::STRING AS col_4017, | e'\u00C0':::STRING AS col_4018, | (-6.201746856518076491):::DECIMAL AS col_4019, | tab_1393.col3_14 AS col_4020, | (9.9:::DECIMAL::DECIMAL / 1:::INT8::INT8)::DECIMAL AS col_4021 | FROM | defaultdb.public.table3@[0] AS tab_1393 | LEFT JOIN defaultdb.public.table3@[0] AS tab_1394 ON | (tab_1393.col3_0) = (tab_1394.col3_3) AND (tab_1393.crdb_internal_mvcc_timestamp) >= (tab_1394.col3_12) | ORDER BY | tab_1393.col3_8 DESC, tab_1394.col3_13 ASC Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=all.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20197
2.0
roachtest: unoptimized-query-oracle/disable-rules=all failed - roachtest.unoptimized-query-oracle/disable-rules=all [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6775433?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6775433?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=all) on master @ [0eaeeb773474716753781289788fdd087fb9b166](https://github.com/cockroachdb/cockroach/commits/0eaeeb773474716753781289788fdd087fb9b166): ``` |    }, ""), |    strings.Join({ |    ... // 769 identical bytes |    "0C18A117E67086FEEC1EABF114477C1E4C15123643403EFF3C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |    strings.Join({ |    ... // 953 identical bytes |    "5C150B766818ABAF0C120518B2F2EAEFE414CD73A1320CEE8C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |    strings.Join({ |    ... // 2965 identical bytes |    "3C1E498DCAF61A4E141C0587B873586F1C1C019D78E4F37D2C1,NULL,1987-12", |    "-04 01:57:20.000745 +0000 +0000,!,À,-6.201746856518076491,0,9.9", | +  "000000000000000000", |    }, ""), |   } | sql: SELECT | '1983-07-24 03:35:34.000493':::TIMESTAMP AS col_4000, | tab_1393.col3_8 AS col_4001, | (-1.0717353820800781):::FLOAT8 AS col_4002, | 0:::OID AS col_4003, | tab_1394.col3_13 AS col_4004, | NULL AS col_4005, | tab_1393.col3_2 AS col_4006, | tab_1394.col3_1 AS col_4007, | (-7318860970.773327295):::DECIMAL AS col_4008, | tab_1393.col3_11 AS col_4009, | tab_1393.col3_7 AS col_4010, | tab_1394.col3_0 AS col_4011, | tab_1394.col3_10 AS col_4012, | tab_1393.col3_5 AS col_4013, | tab_1393.col3_0 AS col_4014, | tab_1394.col3_14 AS col_4015, | '1987-12-04 01:57:20.000745':::TIMESTAMP AS col_4016, | '!':::STRING AS col_4017, | e'\u00C0':::STRING AS col_4018, | (-6.201746856518076491):::DECIMAL AS col_4019, | tab_1393.col3_14 AS col_4020, | (9.9:::DECIMAL::DECIMAL / 1:::INT8::INT8)::DECIMAL AS col_4021 | FROM | defaultdb.public.table3@[0] AS tab_1393 | LEFT JOIN defaultdb.public.table3@[0] AS tab_1394 ON | (tab_1393.col3_0) = (tab_1394.col3_3) AND (tab_1393.crdb_internal_mvcc_timestamp) >= (tab_1394.col3_12) | ORDER BY | tab_1393.col3_8 DESC, tab_1394.col3_13 ASC Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=all.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20197
test
roachtest unoptimized query oracle disable rules all failed roachtest unoptimized query oracle disable rules all with on master       strings join    identical bytes    null    à         strings join    identical bytes    null    à         strings join    identical bytes    null    à         sql select timestamp as col tab as col as col oid as col tab as col null as col tab as col tab as col decimal as col tab as col tab as col tab as col tab as col tab as col tab as col tab as col timestamp as col string as col e string as col decimal as col tab as col decimal decimal decimal as col from defaultdb public as tab left join defaultdb public as tab on tab tab and tab crdb internal mvcc timestamp tab order by tab desc tab asc error types withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql queries jira issue crdb
1
63,985
6,889,853,461
IssuesEvent
2017-11-22 11:52:55
RIOT-OS/RIOT
https://api.github.com/repos/RIOT-OS/RIOT
opened
crypto/ccm: bugs in the implementation of CCM mode
bug crypto tests
This issue summarize all the issues I have found when trying to add tests for #6706 for `sys/crypto/modes/ccm.c` Issues shown in the PR: * Upper bound for `auth_data_len` should be `0xFEFF`. https://tools.ietf.org/html/rfc3610 2.2 page * Wrong check for input plaintext message length against len encoding. * buffer overflow if auth_data_len is larger than 26 Problems found when adding tests: * In tests 'len_encoding' and 'mac_length' are hardwritten. * Tests do not support `input_len` >= 256 * ccm code will break if length_encoding >= 4 (even with PR fix because of overflow). * Code is not saving enough bytes when saving plaintext_length (off by one limit) * CCM is broken with `input_len` >= 256 (so broken even when fixing previous point) - [ ] Refactor test to allow running other tests - [ ] Fix and add test for auth_data_len upper bound - [ ] Fix and add test for input plaintext message length check - [ ] Fix and add test for auth_data_len > 26 - [ ] Fix and test for input_len >= 256 - [ ] Add all RFC/NIST tests
1.0
crypto/ccm: bugs in the implementation of CCM mode - This issue summarize all the issues I have found when trying to add tests for #6706 for `sys/crypto/modes/ccm.c` Issues shown in the PR: * Upper bound for `auth_data_len` should be `0xFEFF`. https://tools.ietf.org/html/rfc3610 2.2 page * Wrong check for input plaintext message length against len encoding. * buffer overflow if auth_data_len is larger than 26 Problems found when adding tests: * In tests 'len_encoding' and 'mac_length' are hardwritten. * Tests do not support `input_len` >= 256 * ccm code will break if length_encoding >= 4 (even with PR fix because of overflow). * Code is not saving enough bytes when saving plaintext_length (off by one limit) * CCM is broken with `input_len` >= 256 (so broken even when fixing previous point) - [ ] Refactor test to allow running other tests - [ ] Fix and add test for auth_data_len upper bound - [ ] Fix and add test for input plaintext message length check - [ ] Fix and add test for auth_data_len > 26 - [ ] Fix and test for input_len >= 256 - [ ] Add all RFC/NIST tests
test
crypto ccm bugs in the implementation of ccm mode this issue summarize all the issues i have found when trying to add tests for for sys crypto modes ccm c issues shown in the pr upper bound for auth data len should be page wrong check for input plaintext message length against len encoding buffer overflow if auth data len is larger than problems found when adding tests in tests len encoding and mac length are hardwritten tests do not support input len ccm code will break if length encoding even with pr fix because of overflow code is not saving enough bytes when saving plaintext length off by one limit ccm is broken with input len so broken even when fixing previous point refactor test to allow running other tests fix and add test for auth data len upper bound fix and add test for input plaintext message length check fix and add test for auth data len fix and test for input len add all rfc nist tests
1
343,433
30,665,588,758
IssuesEvent
2023-07-25 18:01:23
opensearch-project/alerting-dashboards-plugin
https://api.github.com/repos/opensearch-project/alerting-dashboards-plugin
closed
[AUTOCUT] Integration Test failed for alertingDashboards: 2.9.0 rpm distribution
untriaged autocut integ-test-failure v2.9.0
The integration test failed at distribution level for component alertingDashboards<br>Version: 2.9.0<br>Distribution: rpm<br>Architecture: x64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test-opensearch-dashboards/3690/display/redirect<br><br> * Steps to reproduce: See https://github.com/opensearch-project/opensearch-build/tree/main/src/test_workflow#integration-tests<br>* See all log files:<br> - [With security](https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.9.0/6388/linux/x64/rpm/test-results/3690/integ-test/alertingDashboards/with-security/alertingDashboards.yml) (if applicable)<br> - [Without security](https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.9.0/6388/linux/x64/rpm/test-results/3690/integ-test/alertingDashboards/without-security/alertingDashboards.yml) (if applicable)<br><br>
1.0
[AUTOCUT] Integration Test failed for alertingDashboards: 2.9.0 rpm distribution - The integration test failed at distribution level for component alertingDashboards<br>Version: 2.9.0<br>Distribution: rpm<br>Architecture: x64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test-opensearch-dashboards/3690/display/redirect<br><br> * Steps to reproduce: See https://github.com/opensearch-project/opensearch-build/tree/main/src/test_workflow#integration-tests<br>* See all log files:<br> - [With security](https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.9.0/6388/linux/x64/rpm/test-results/3690/integ-test/alertingDashboards/with-security/alertingDashboards.yml) (if applicable)<br> - [Without security](https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.9.0/6388/linux/x64/rpm/test-results/3690/integ-test/alertingDashboards/without-security/alertingDashboards.yml) (if applicable)<br><br>
test
integration test failed for alertingdashboards rpm distribution the integration test failed at distribution level for component alertingdashboards version distribution rpm architecture platform linux please check the logs steps to reproduce see see all log files if applicable if applicable
1
33,459
15,953,281,303
IssuesEvent
2021-04-15 12:15:22
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
`tensor.mode()` takes extraordinarily long for sizes larger than `65536`
module: cuda module: performance module: sorting and selection triaged
## 🐛 Bug See title. The operation below takes around 20 seconds. On cpu(), or with batch size smaller than 65336, it's less than a second. ## To Reproduce ``` a = torch.randint(0, 2, size=(65537, 50)).cuda() a.mode(-1) ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment - PyTorch Version (e.g., 1.0): 1.3 - OS (e.g., Linux): Linux - How you installed PyTorch (`conda`, `pip`, source): pip - Python version: 3.6 ## Additional context <!-- Add any other context about the problem here. --> cc @ngimel @VitalyFedyunin @heitorschueroff
True
`tensor.mode()` takes extraordinarily long for sizes larger than `65536` - ## 🐛 Bug See title. The operation below takes around 20 seconds. On cpu(), or with batch size smaller than 65336, it's less than a second. ## To Reproduce ``` a = torch.randint(0, 2, size=(65537, 50)).cuda() a.mode(-1) ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment - PyTorch Version (e.g., 1.0): 1.3 - OS (e.g., Linux): Linux - How you installed PyTorch (`conda`, `pip`, source): pip - Python version: 3.6 ## Additional context <!-- Add any other context about the problem here. --> cc @ngimel @VitalyFedyunin @heitorschueroff
non_test
tensor mode takes extraordinarily long for sizes larger than 🐛 bug see title the operation below takes around seconds on cpu or with batch size smaller than it s less than a second to reproduce a torch randint size cuda a mode expected behavior environment pytorch version e g os e g linux linux how you installed pytorch conda pip source pip python version additional context cc ngimel vitalyfedyunin heitorschueroff
0
436,762
30,568,778,445
IssuesEvent
2023-07-20 20:05:01
EdgePi-Cloud/edgepi-python-sdk
https://api.github.com/repos/EdgePi-Cloud/edgepi-python-sdk
closed
PWM Usage Doc
documentation PWM
- enabling PWM throws invalid argument error when the PWM device is not configured - set_config() method needs to be executed first before enabling the PWM - both frequency and duty cycle need to be configured
1.0
PWM Usage Doc - - enabling PWM throws invalid argument error when the PWM device is not configured - set_config() method needs to be executed first before enabling the PWM - both frequency and duty cycle need to be configured
non_test
pwm usage doc enabling pwm throws invalid argument error when the pwm device is not configured set config method needs to be executed first before enabling the pwm both frequency and duty cycle need to be configured
0
296,613
25,563,116,031
IssuesEvent
2022-11-30 12:21:55
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
com.hazelcast.internal.serialization.impl.CompatibilitySerializationServiceTest fails with little-endian [HZ-1824]
Team: Client Team: Core Type: Test-Failure Source: Internal to-jira constant-test-failure
_master_ (commit d11872d0d44b077ab5367bf2ef3218063cd06842) Failed on with little-endian (Oracle JDK 8): https://jenkins.hazelcast.com/job/Hazelcast-master-little-endian/228/testReport/com.hazelcast.internal.serialization.impl/CompatibilitySerializationServiceTest/ <details><summary>Stacktrace:</summary> ``` org.junit.ComparisonFailure: expected:<... -1, -1, -1, -2, 0, [0, 0, 0, 52, 99, 111, 109, 46, 104, 97, 122, 101, 108, 99, 97, 115, 116, 46, 105, 110, 116, 101, 114, 110, 97, 108, 46, 115, 101, 114, 105, 97, 108, 105, 122, 97, 116, 105, 111, 110, 46, 105, 109, 112, 108, 46, 84, 101, 115, 116, 79, 98, 106, 101, 99, 116, 1, 42, 0, 100, 0, 42, 0, 0, 0, 42, 0, 0, 0, 0, 0, 0, 0, 42, 64, -122, 102, 102, 64, 16, -52, -52, -52, -52, -52, -51, 0, 0, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -1, -1, -1, -26, 0, 0, 0, 1, -1, -1, -1, -11, 0, 0, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -1, -1, -1, -100, -84, -19, 0, 5, 115, 114, 0, 17, 106, 97, 118, 97, 46, 117, 116, 105, 108, 46, 72, 97, 115, 104, 83, 101, 116, -70, 68, -123, -107, -106, -72, -73, 52, 3, 0, 0, 120, 112, 119, 12, 0, 0, 0, 2, 63, 64, 0, 0, 0, 0, 0, 1, 116, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, 120, -1, -1, -1, -100], -84, -19, 0, 5, 11...> but was:<... -1, -1, -1, -2, 0, [52, 0, 0, 0, 99, 111, 109, 46, 104, 97, 122, 101, 108, 99, 97, 115, 116, 46, 105, 110, 116, 101, 114, 110, 97, 108, 46, 115, 101, 114, 105, 97, 108, 105, 122, 97, 116, 105, 111, 110, 46, 105, 109, 112, 108, 46, 84, 101, 115, 116, 79, 98, 106, 101, 99, 116, 1, 42, 100, 0, 42, 0, 42, 0, 0, 0, 42, 0, 0, 0, 0, 0, 0, 0, 102, 102, -122, 64, -51, -52, -52, -52, -52, -52, 16, 64, 11, 0, 0, 0, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -26, -1, -1, -1, 1, 0, 0, 0, -11, -1, -1, -1, 11, 0, 0, 0, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -100, -1, -1, -1, -84, -19, 0, 5, 115, 114, 0, 17, 106, 97, 118, 97, 46, 117, 116, 105, 108, 46, 72, 97, 115, 104, 83, 101, 116, -70, 68, -123, -107, -106, -72, -73, 52, 3, 0, 0, 120, 112, 119, 12, 0, 0, 0, 2, 63, 64, 0, 0, 0, 0, 0, 1, 116, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, 120, -100, -1, -1, -1], -84, -19, 0, 5, 11...> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at com.hazelcast.internal.serialization.impl.CompatibilitySerializationServiceTest.testSerializeObjectAsHz3Data(CompatibilitySerializationServiceTest.java:89) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:750) ``` </details> Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-master-little-endian/228/
2.0
com.hazelcast.internal.serialization.impl.CompatibilitySerializationServiceTest fails with little-endian [HZ-1824] - _master_ (commit d11872d0d44b077ab5367bf2ef3218063cd06842) Failed on with little-endian (Oracle JDK 8): https://jenkins.hazelcast.com/job/Hazelcast-master-little-endian/228/testReport/com.hazelcast.internal.serialization.impl/CompatibilitySerializationServiceTest/ <details><summary>Stacktrace:</summary> ``` org.junit.ComparisonFailure: expected:<... -1, -1, -1, -2, 0, [0, 0, 0, 52, 99, 111, 109, 46, 104, 97, 122, 101, 108, 99, 97, 115, 116, 46, 105, 110, 116, 101, 114, 110, 97, 108, 46, 115, 101, 114, 105, 97, 108, 105, 122, 97, 116, 105, 111, 110, 46, 105, 109, 112, 108, 46, 84, 101, 115, 116, 79, 98, 106, 101, 99, 116, 1, 42, 0, 100, 0, 42, 0, 0, 0, 42, 0, 0, 0, 0, 0, 0, 0, 42, 64, -122, 102, 102, 64, 16, -52, -52, -52, -52, -52, -51, 0, 0, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -1, -1, -1, -26, 0, 0, 0, 1, -1, -1, -1, -11, 0, 0, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -1, -1, -1, -100, -84, -19, 0, 5, 115, 114, 0, 17, 106, 97, 118, 97, 46, 117, 116, 105, 108, 46, 72, 97, 115, 104, 83, 101, 116, -70, 68, -123, -107, -106, -72, -73, 52, 3, 0, 0, 120, 112, 119, 12, 0, 0, 0, 2, 63, 64, 0, 0, 0, 0, 0, 1, 116, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, 120, -1, -1, -1, -100], -84, -19, 0, 5, 11...> but was:<... -1, -1, -1, -2, 0, [52, 0, 0, 0, 99, 111, 109, 46, 104, 97, 122, 101, 108, 99, 97, 115, 116, 46, 105, 110, 116, 101, 114, 110, 97, 108, 46, 115, 101, 114, 105, 97, 108, 105, 122, 97, 116, 105, 111, 110, 46, 105, 109, 112, 108, 46, 84, 101, 115, 116, 79, 98, 106, 101, 99, 116, 1, 42, 100, 0, 42, 0, 42, 0, 0, 0, 42, 0, 0, 0, 0, 0, 0, 0, 102, 102, -122, 64, -51, -52, -52, -52, -52, -52, 16, 64, 11, 0, 0, 0, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -26, -1, -1, -1, 1, 0, 0, 0, -11, -1, -1, -1, 11, 0, 0, 0, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, -100, -1, -1, -1, -84, -19, 0, 5, 115, 114, 0, 17, 106, 97, 118, 97, 46, 117, 116, 105, 108, 46, 72, 97, 115, 104, 83, 101, 116, -70, 68, -123, -107, -106, -72, -73, 52, 3, 0, 0, 120, 112, 119, 12, 0, 0, 0, 2, 63, 64, 0, 0, 0, 0, 0, 1, 116, 0, 11, 104, 101, 108, 108, 111, 32, 119, 111, 114, 108, 100, 120, -100, -1, -1, -1], -84, -19, 0, 5, 11...> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at com.hazelcast.internal.serialization.impl.CompatibilitySerializationServiceTest.testSerializeObjectAsHz3Data(CompatibilitySerializationServiceTest.java:89) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:750) ``` </details> Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-master-little-endian/228/
test
com hazelcast internal serialization impl compatibilityserializationservicetest fails with little endian master commit failed on with little endian oracle jdk stacktrace org junit comparisonfailure expected but was at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at com hazelcast internal serialization impl compatibilityserializationservicetest compatibilityserializationservicetest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java standard output can be found here
1
806,089
29,799,704,149
IssuesEvent
2023-06-16 07:07:13
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
storage: many tests failed
type: bug api: storage priority: p1 flakybot: issue
Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * TestIntegration_BucketLifecycle/grpc * TestIntegration_BucketUpdate/grpc * TestIntegration_BucketPolicyOnly/grpc (#7401) * TestIntegration_UniformBucketLevelAccess/grpc (#7405) * TestIntegration_PublicAccessPrevention/grpc (#6964) * TestIntegration_Autoclass/grpc * TestIntegration_ConditionalDelete/grpc * TestIntegration_ObjectsRangeReader/grpc/offset_and_read_till_end * TestIntegration_ObjectReadChunksGRPC/grpc * TestIntegration_MultiMessageWriteGRPC/grpc (#5372) * TestIntegration_MultiChunkWrite/grpc (#7033) * TestIntegration_ConditionalDownload/grpc * TestIntegration_ObjectIteration/grpc (#8073) * TestIntegration_ObjectChecksums/grpc (#8050) * TestIntegration_ObjectCompose/grpc * TestIntegration_SignedURL/grpc * TestIntegration_WriterContentType/grpc * TestIntegration_WriterChunksize/grpc/multiple_of_256kib * TestIntegration_ZeroSizedObject/grpc (#5786) * TestIntegration_RequesterPaysOwner/grpc/cannot_use_someone_else's_project_for_billing (#7483) * TestIntegration_RequesterPaysNonOwner/grpc/user_is_not_an_Editor_on_UserProject (#7484) * TestIntegration_PublicObject/grpc * TestIntegration_UpdateCORS/grpc/do_not_set_to_keep_existing_policies (#7465) * TestIntegration_UpdateDefaultEventBasedHold/grpc * TestIntegration_UpdateEventBasedHold/grpc (#8067) * TestIntegration_UpdateTemporaryHold/grpc * TestIntegration_UpdateRetentionExpirationTime/grpc (#5024) * TestIntegration_CustomTime/grpc * TestIntegration_DeleteObjectInBucketWithRetentionPolicy/grpc * TestIntegration_KMS/grpc (#7993) * TestIntegration_Reader/grpc * TestIntegration_ReaderAttrs/grpc (#8116) ----- commit: 28aa098866f62b04728704d26d16b68d5a034ec7 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/76715a0f-fb15-49ca-af46-0072e7a5c9bb), [Sponge](http://sponge2/76715a0f-fb15-49ca-af46-0072e7a5c9bb) status: failed
1.0
storage: many tests failed - Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * TestIntegration_BucketLifecycle/grpc * TestIntegration_BucketUpdate/grpc * TestIntegration_BucketPolicyOnly/grpc (#7401) * TestIntegration_UniformBucketLevelAccess/grpc (#7405) * TestIntegration_PublicAccessPrevention/grpc (#6964) * TestIntegration_Autoclass/grpc * TestIntegration_ConditionalDelete/grpc * TestIntegration_ObjectsRangeReader/grpc/offset_and_read_till_end * TestIntegration_ObjectReadChunksGRPC/grpc * TestIntegration_MultiMessageWriteGRPC/grpc (#5372) * TestIntegration_MultiChunkWrite/grpc (#7033) * TestIntegration_ConditionalDownload/grpc * TestIntegration_ObjectIteration/grpc (#8073) * TestIntegration_ObjectChecksums/grpc (#8050) * TestIntegration_ObjectCompose/grpc * TestIntegration_SignedURL/grpc * TestIntegration_WriterContentType/grpc * TestIntegration_WriterChunksize/grpc/multiple_of_256kib * TestIntegration_ZeroSizedObject/grpc (#5786) * TestIntegration_RequesterPaysOwner/grpc/cannot_use_someone_else's_project_for_billing (#7483) * TestIntegration_RequesterPaysNonOwner/grpc/user_is_not_an_Editor_on_UserProject (#7484) * TestIntegration_PublicObject/grpc * TestIntegration_UpdateCORS/grpc/do_not_set_to_keep_existing_policies (#7465) * TestIntegration_UpdateDefaultEventBasedHold/grpc * TestIntegration_UpdateEventBasedHold/grpc (#8067) * TestIntegration_UpdateTemporaryHold/grpc * TestIntegration_UpdateRetentionExpirationTime/grpc (#5024) * TestIntegration_CustomTime/grpc * TestIntegration_DeleteObjectInBucketWithRetentionPolicy/grpc * TestIntegration_KMS/grpc (#7993) * TestIntegration_Reader/grpc * TestIntegration_ReaderAttrs/grpc (#8116) ----- commit: 28aa098866f62b04728704d26d16b68d5a034ec7 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/76715a0f-fb15-49ca-af46-0072e7a5c9bb), [Sponge](http://sponge2/76715a0f-fb15-49ca-af46-0072e7a5c9bb) status: failed
non_test
storage many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed testintegration bucketlifecycle grpc testintegration bucketupdate grpc testintegration bucketpolicyonly grpc testintegration uniformbucketlevelaccess grpc testintegration publicaccessprevention grpc testintegration autoclass grpc testintegration conditionaldelete grpc testintegration objectsrangereader grpc offset and read till end testintegration objectreadchunksgrpc grpc testintegration multimessagewritegrpc grpc testintegration multichunkwrite grpc testintegration conditionaldownload grpc testintegration objectiteration grpc testintegration objectchecksums grpc testintegration objectcompose grpc testintegration signedurl grpc testintegration writercontenttype grpc testintegration writerchunksize grpc multiple of testintegration zerosizedobject grpc testintegration requesterpaysowner grpc cannot use someone else s project for billing testintegration requesterpaysnonowner grpc user is not an editor on userproject testintegration publicobject grpc testintegration updatecors grpc do not set to keep existing policies testintegration updatedefaulteventbasedhold grpc testintegration updateeventbasedhold grpc testintegration updatetemporaryhold grpc testintegration updateretentionexpirationtime grpc testintegration customtime grpc testintegration deleteobjectinbucketwithretentionpolicy grpc testintegration kms grpc testintegration reader grpc testintegration readerattrs grpc commit buildurl status failed
0
1,124
2,532,268,947
IssuesEvent
2015-01-23 15:01:14
AAndharia/ZIMS-School-Mgmt
https://api.github.com/repos/AAndharia/ZIMS-School-Mgmt
closed
Add cancel button
Enhancement Tested & Verified
On all add/edit pages add cancel button ... that will take user back to list page ... and hide reset button for now... just hide do not remove from coding
1.0
Add cancel button - On all add/edit pages add cancel button ... that will take user back to list page ... and hide reset button for now... just hide do not remove from coding
test
add cancel button on all add edit pages add cancel button that will take user back to list page and hide reset button for now just hide do not remove from coding
1
44,008
5,580,677,155
IssuesEvent
2017-03-28 17:08:59
piwik/piwik
https://api.github.com/repos/piwik/piwik
opened
Some Overlay UI tests are failing
c: Tests & QA not-in-changelog Regression
Following UI tests have regressed recently and are now failing: * Overlay_page_new_links.png * Overlay_row_evolution.png * Overlay_transitions.png Example: http://builds-artifacts.piwik.org/piwik/piwik/Findus23-icons-as-a-submodule/22704/ We've seens this particular regression in the Overlay UI tests several times over the last few months/years. It should be very interesting to find out what is the cause this time.
1.0
Some Overlay UI tests are failing - Following UI tests have regressed recently and are now failing: * Overlay_page_new_links.png * Overlay_row_evolution.png * Overlay_transitions.png Example: http://builds-artifacts.piwik.org/piwik/piwik/Findus23-icons-as-a-submodule/22704/ We've seens this particular regression in the Overlay UI tests several times over the last few months/years. It should be very interesting to find out what is the cause this time.
test
some overlay ui tests are failing following ui tests have regressed recently and are now failing overlay page new links png overlay row evolution png overlay transitions png example we ve seens this particular regression in the overlay ui tests several times over the last few months years it should be very interesting to find out what is the cause this time
1
291,278
8,922,724,624
IssuesEvent
2019-01-21 13:49:22
phonetworks/graphjs-server
https://api.github.com/repos/phonetworks/graphjs-server
opened
Transform calls to more HTTP verbs
high priority enhancement
like POST, PATCH and rename all public functions accordingly. We already have one POST; see function post of BlogController in master branch. This will show you how it works with this library. Todo list should be as follows - [ ] List all public verbs - [ ] Suggest a version 2 with more verb types - [ ] Implement
1.0
Transform calls to more HTTP verbs - like POST, PATCH and rename all public functions accordingly. We already have one POST; see function post of BlogController in master branch. This will show you how it works with this library. Todo list should be as follows - [ ] List all public verbs - [ ] Suggest a version 2 with more verb types - [ ] Implement
non_test
transform calls to more http verbs like post patch and rename all public functions accordingly we already have one post see function post of blogcontroller in master branch this will show you how it works with this library todo list should be as follows list all public verbs suggest a version with more verb types implement
0
273,668
23,776,097,570
IssuesEvent
2022-09-01 21:06:28
Princeton-CDH/geniza
https://api.github.com/repos/Princeton-CDH/geniza
closed
As a content editor, I want new and revised transcriptions available for search immediately so that changes and new content are all available to all site users.
🗜️ awaiting testing
## testing notes - [x] add and edit a handful of transcriptions as described in #1002 and #913 - [x] go to the public site search and search for the text of the transcriptions you added/edited, and confirm it brings up the document(s) you transcribed successfully ## dev notes - [x] revise document index data to pull transcription content from annotations - [x] add a new signal handler to reindex documents when annotations associated with them are changed or added (maybe via footnote / scholarship record?) - [x] signal handler to reindex document when annotation changes but footnote does not
1.0
As a content editor, I want new and revised transcriptions available for search immediately so that changes and new content are all available to all site users. - ## testing notes - [x] add and edit a handful of transcriptions as described in #1002 and #913 - [x] go to the public site search and search for the text of the transcriptions you added/edited, and confirm it brings up the document(s) you transcribed successfully ## dev notes - [x] revise document index data to pull transcription content from annotations - [x] add a new signal handler to reindex documents when annotations associated with them are changed or added (maybe via footnote / scholarship record?) - [x] signal handler to reindex document when annotation changes but footnote does not
test
as a content editor i want new and revised transcriptions available for search immediately so that changes and new content are all available to all site users testing notes add and edit a handful of transcriptions as described in and go to the public site search and search for the text of the transcriptions you added edited and confirm it brings up the document s you transcribed successfully dev notes revise document index data to pull transcription content from annotations add a new signal handler to reindex documents when annotations associated with them are changed or added maybe via footnote scholarship record signal handler to reindex document when annotation changes but footnote does not
1
132,925
28,437,456,000
IssuesEvent
2023-04-15 13:42:34
HYF-Class20/agile-development-group2-loruki
https://api.github.com/repos/HYF-Class20/agile-development-group2-loruki
closed
Code : HTML : Creating language and footer
documentation Code HTML
We need a branch for our language and footer part Inside that branch we need: Inside that branch we need: - [ ] language and footer must be inside Body - [ ] create a section for language and footer - [ ] create paragraphs divs language and footer
1.0
Code : HTML : Creating language and footer - We need a branch for our language and footer part Inside that branch we need: Inside that branch we need: - [ ] language and footer must be inside Body - [ ] create a section for language and footer - [ ] create paragraphs divs language and footer
non_test
code html creating language and footer we need a branch for our language and footer part inside that branch we need inside that branch we need language and footer must be inside body create a section for language and footer create paragraphs divs language and footer
0