Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
27,608
29,801,972,350
IssuesEvent
2023-06-16 08:47:59
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Unwarranted value incremented in text boxes on focus
bug topic:editor usability regression
### Godot version 4.1 beta1, 4.1.beta2 ### System information Windows 11 ### Issue description Whenever you try to select an editable field like Scale.X and Scale.Y. The value is changed as soon as the focus is on. This is random behavior but consistently reproducible throughout the editor. https://github.com/godotengine/godot/assets/84369217/2773c297-bf4b-4dd7-86f7-0c837b8d4a12 **It is not reproducible in 4.0.3** ### Steps to reproduce Reproducible in most of the editable text fields with floating point values. It is random but consistent ### Minimal reproduction project N.A
True
Unwarranted value incremented in text boxes on focus - ### Godot version 4.1 beta1, 4.1.beta2 ### System information Windows 11 ### Issue description Whenever you try to select an editable field like Scale.X and Scale.Y. The value is changed as soon as the focus is on. This is random behavior but consistently reproducible throughout the editor. https://github.com/godotengine/godot/assets/84369217/2773c297-bf4b-4dd7-86f7-0c837b8d4a12 **It is not reproducible in 4.0.3** ### Steps to reproduce Reproducible in most of the editable text fields with floating point values. It is random but consistent ### Minimal reproduction project N.A
non_process
unwarranted value incremented in text boxes on focus godot version system information windows issue description whenever you try to select an editable field like scale x and scale y the value is changed as soon as the focus is on this is random behavior but consistently reproducible throughout the editor it is not reproducible in steps to reproduce reproducible in most of the editable text fields with floating point values it is random but consistent minimal reproduction project n a
0
8,414
11,580,019,686
IssuesEvent
2020-02-21 19:13:08
googleapis/python-pubsub
https://api.github.com/repos/googleapis/python-pubsub
closed
PubSub: Add more system tests covering various RBAC-related scenarios
api: pubsub testing type: process
A [regression](https://github.com/googleapis/google-cloud-python/issues/9339) introduced not so long ago could have been prevented if we had more RBAC-related system tests. A system test covering the fix was added in googleapis/google-cloud-python#9507. Since PubSub backend defines several different [user roles](https://cloud.google.com/pubsub/docs/access-control#roles), system tests must be added to cover at least the most common use case scenarios that can be affected by the user role used in the session.
1.0
PubSub: Add more system tests covering various RBAC-related scenarios - A [regression](https://github.com/googleapis/google-cloud-python/issues/9339) introduced not so long ago could have been prevented if we had more RBAC-related system tests. A system test covering the fix was added in googleapis/google-cloud-python#9507. Since PubSub backend defines several different [user roles](https://cloud.google.com/pubsub/docs/access-control#roles), system tests must be added to cover at least the most common use case scenarios that can be affected by the user role used in the session.
process
pubsub add more system tests covering various rbac related scenarios a introduced not so long ago could have been prevented if we had more rbac related system tests a system test covering the fix was added in googleapis google cloud python since pubsub backend defines several different system tests must be added to cover at least the most common use case scenarios that can be affected by the user role used in the session
1
21,412
29,351,206,474
IssuesEvent
2023-05-27 00:34:49
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Product Designer na Coodesh
SALVADOR UX REQUISITOS REMOTO PROCESSOS GITHUB SEGURANÇA Figma UMA QUALIDADE VENDAS BARREIRAS MANUTENÇÃO USABILIDADE PRODUCT MANAGER UI B2B Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-designer-193018325?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O nosso time de Produto está em busca de um <strong>Product Designer</strong> que consiga compreender o contexto do negócio da Onisys e a realidade dos usuários. Dessa forma, pode criar soluções de produtos intuitivas que facilitem a jornada dos nossos clientes, desenhando experiências e conduzindo projetos de design de ponta a ponta junto aos times de Produto e Tecnologia.</p> <p>Se você tem forte visão analítica, entende de comportamento humano, tecnologia, é apaixonado por UX|UI e consegue desenvolver soluções que combinam os objetivos do negócio com a melhor experiência para o usuário, esta vaga pode ser sua!</p> <p>Suas responsabilidades:</p> <ul> <li>Atuar ativamente no desenho de arquitetura dos produtos, construir protótipos navegáveis de alta e baixa fidelidade, mapas de jornada, interfaces e entregáveis focados na experiência do usuário.</li> <li>Balancear necessidades do usuário com objetivos do negócio, atuando em parceria com os times de Produto Engenharia, Marketing e Vendas para a entrega de novos produtos e evolução dos já existentes.</li> <li>Participar junto ao Product Manager na definição dos processos de produto, removendo barreiras técnicas e provendo visão estratégica de design e usabilidade de curto e longo prazo para alcançar os objetivos do produto.</li> <li>Desenhar, conduzir, analisar e reportar resultados de pesquisas centradas no usuário e testes de usabilidade, assim como conduzir pesquisas quantitativas e cruzar resultados.</li> <li>Analisar métricas e entregar soluções a partir de dados e relatórios recebidos do time de Produto.</li> <li>Participar ativamente na criação, aplicação e manutenção consistente do sistema de design em todos os produtos, garantindo eficiência e velocidade na criação das interfaces.</li> </ul> <p></p> ## Onisys: <p>Plataforma de gestão de Segurança no Transporte. Cuidando da segurança do maior ativo da sua transportadora: o MOTORISTA</p> <p>A plataforma Onisys possui dois planos que conseguem atender a todos os tipos de empresas de transporte, ou empresas que tenham o transporte como parte de sua operação: Onisys Básico e Onisys Safe.</p><a href='https://coodesh.com/empresas/onisys'>Veja mais no site</a> ## Habilidades: - UX/UI - Figma - Design UX ## Local: 100% Remoto ## Requisitos: - Domínio em design de interação, experiência do usuário e interfaces de alta qualidade; - Domínio de metodologias de design centrado no usuário; - Experiência com ferramentas e metodologias de pesquisa com usuários; - Capacidade de executar um projeto de design de ponta a ponta; - Noções de processos de desenvolvimento de produto e metodologia ágil; - Excelente comunicação e capacidade de transmitir ideias de forma clara para pessoas de diferentes áreas de expertise; - Experiência com produtos B2B. ## Benefícios: - Horários Flexíveis; - Vale alimentação; - Plano de saúde integral; - Plano Odontológico Integral. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Designer na Onisys](https://coodesh.com/vagas/product-designer-193018325?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime CLT #### Categoria Design/UX
1.0
[Remoto] Product Designer na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-designer-193018325?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O nosso time de Produto está em busca de um <strong>Product Designer</strong> que consiga compreender o contexto do negócio da Onisys e a realidade dos usuários. Dessa forma, pode criar soluções de produtos intuitivas que facilitem a jornada dos nossos clientes, desenhando experiências e conduzindo projetos de design de ponta a ponta junto aos times de Produto e Tecnologia.</p> <p>Se você tem forte visão analítica, entende de comportamento humano, tecnologia, é apaixonado por UX|UI e consegue desenvolver soluções que combinam os objetivos do negócio com a melhor experiência para o usuário, esta vaga pode ser sua!</p> <p>Suas responsabilidades:</p> <ul> <li>Atuar ativamente no desenho de arquitetura dos produtos, construir protótipos navegáveis de alta e baixa fidelidade, mapas de jornada, interfaces e entregáveis focados na experiência do usuário.</li> <li>Balancear necessidades do usuário com objetivos do negócio, atuando em parceria com os times de Produto Engenharia, Marketing e Vendas para a entrega de novos produtos e evolução dos já existentes.</li> <li>Participar junto ao Product Manager na definição dos processos de produto, removendo barreiras técnicas e provendo visão estratégica de design e usabilidade de curto e longo prazo para alcançar os objetivos do produto.</li> <li>Desenhar, conduzir, analisar e reportar resultados de pesquisas centradas no usuário e testes de usabilidade, assim como conduzir pesquisas quantitativas e cruzar resultados.</li> <li>Analisar métricas e entregar soluções a partir de dados e relatórios recebidos do time de Produto.</li> <li>Participar ativamente na criação, aplicação e manutenção consistente do sistema de design em todos os produtos, garantindo eficiência e velocidade na criação das interfaces.</li> </ul> <p></p> ## Onisys: <p>Plataforma de gestão de Segurança no Transporte. Cuidando da segurança do maior ativo da sua transportadora: o MOTORISTA</p> <p>A plataforma Onisys possui dois planos que conseguem atender a todos os tipos de empresas de transporte, ou empresas que tenham o transporte como parte de sua operação: Onisys Básico e Onisys Safe.</p><a href='https://coodesh.com/empresas/onisys'>Veja mais no site</a> ## Habilidades: - UX/UI - Figma - Design UX ## Local: 100% Remoto ## Requisitos: - Domínio em design de interação, experiência do usuário e interfaces de alta qualidade; - Domínio de metodologias de design centrado no usuário; - Experiência com ferramentas e metodologias de pesquisa com usuários; - Capacidade de executar um projeto de design de ponta a ponta; - Noções de processos de desenvolvimento de produto e metodologia ágil; - Excelente comunicação e capacidade de transmitir ideias de forma clara para pessoas de diferentes áreas de expertise; - Experiência com produtos B2B. ## Benefícios: - Horários Flexíveis; - Vale alimentação; - Plano de saúde integral; - Plano Odontológico Integral. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Designer na Onisys](https://coodesh.com/vagas/product-designer-193018325?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime CLT #### Categoria Design/UX
process
product designer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 o nosso time de produto está em busca de um product designer que consiga compreender o contexto do negócio da onisys e a realidade dos usuários dessa forma pode criar soluções de produtos intuitivas que facilitem a jornada dos nossos clientes desenhando experiências e conduzindo projetos de design de ponta a ponta junto aos times de produto e tecnologia se você tem forte visão analítica entende de comportamento humano tecnologia é apaixonado por ux ui e consegue desenvolver soluções que combinam os objetivos do negócio com a melhor experiência para o usuário esta vaga pode ser sua suas responsabilidades atuar ativamente no desenho de arquitetura dos produtos construir protótipos navegáveis de alta e baixa fidelidade mapas de jornada interfaces e entregáveis focados na experiência do usuário balancear necessidades do usuário com objetivos do negócio atuando em parceria com os times de produto engenharia marketing e vendas para a entrega de novos produtos e evolução dos já existentes participar junto ao product manager na definição dos processos de produto removendo barreiras técnicas e provendo visão estratégica de design e usabilidade de curto e longo prazo para alcançar os objetivos do produto desenhar conduzir analisar e reportar resultados de pesquisas centradas no usuário e testes de usabilidade assim como conduzir pesquisas quantitativas e cruzar resultados analisar métricas e entregar soluções a partir de dados e relatórios recebidos do time de produto participar ativamente na criação aplicação e manutenção consistente do sistema de design em todos os produtos garantindo eficiência e velocidade na criação das interfaces onisys plataforma de gestão de segurança no transporte cuidando da segurança do maior ativo da sua transportadora o motorista a plataforma onisys possui dois planos que conseguem atender a todos os tipos de empresas de transporte ou empresas que tenham o transporte como parte de sua operação onisys básico e onisys safe habilidades ux ui figma design ux local remoto requisitos domínio em design de interação experiência do usuário e interfaces de alta qualidade domínio de metodologias de design centrado no usuário experiência com ferramentas e metodologias de pesquisa com usuários capacidade de executar um projeto de design de ponta a ponta noções de processos de desenvolvimento de produto e metodologia ágil excelente comunicação e capacidade de transmitir ideias de forma clara para pessoas de diferentes áreas de expertise experiência com produtos benefícios horários flexíveis vale alimentação plano de saúde integral plano odontológico integral como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime clt categoria design ux
1
550
3,009,369,197
IssuesEvent
2015-07-28 05:32:43
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На бэке главного портала (wf-central) - подвязать вызовы всех сервисов по документам к сохранению истории (Мой журнал)
hi priority In process of testing test
Встроить сохранение истории, при использовании сервисов: 9. Работа с документами 11. DocumentAccess (описание сервисов: https://github.com/e-government-ua/i/blob/test/docs/specification.md ) т.е. в яве(скорей всего в контроллерах вышеперечисленных сервисов) - подобавлять прямой вызов метода "setHistoryEvent" (не через REST, конечно, а напрямую) Так-же ориентироваться на эту концепцию (2. Мой журнал): https://docs.google.com/document/d/1HNA57PO7aC7SJHyWb-CS8OU-w5K1toQtwViyla1VbB0/edit и реализованную задачу: https://github.com/e-government-ua/i/issues/290
1.0
На бэке главного портала (wf-central) - подвязать вызовы всех сервисов по документам к сохранению истории (Мой журнал) - Встроить сохранение истории, при использовании сервисов: 9. Работа с документами 11. DocumentAccess (описание сервисов: https://github.com/e-government-ua/i/blob/test/docs/specification.md ) т.е. в яве(скорей всего в контроллерах вышеперечисленных сервисов) - подобавлять прямой вызов метода "setHistoryEvent" (не через REST, конечно, а напрямую) Так-же ориентироваться на эту концепцию (2. Мой журнал): https://docs.google.com/document/d/1HNA57PO7aC7SJHyWb-CS8OU-w5K1toQtwViyla1VbB0/edit и реализованную задачу: https://github.com/e-government-ua/i/issues/290
process
на бэке главного портала wf central подвязать вызовы всех сервисов по документам к сохранению истории мой журнал встроить сохранение истории при использовании сервисов работа с документами documentaccess описание сервисов т е в яве скорей всего в контроллерах вышеперечисленных сервисов подобавлять прямой вызов метода sethistoryevent не через rest конечно а напрямую так же ориентироваться на эту концепцию мой журнал и реализованную задачу
1
17,832
23,770,923,572
IssuesEvent
2022-09-01 16:13:52
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsoletion: viral assembly intermediate ?????
PomBase obsoletion multi-species process
viral assembly intermediate Specific locations and structures in the virus infected cell involved in assembling new virions. ????? 2 annotations both ISS
1.0
Obsoletion: viral assembly intermediate ????? - viral assembly intermediate Specific locations and structures in the virus infected cell involved in assembling new virions. ????? 2 annotations both ISS
process
obsoletion viral assembly intermediate viral assembly intermediate specific locations and structures in the virus infected cell involved in assembling new virions annotations both iss
1
17,264
23,044,080,065
IssuesEvent
2022-07-23 16:10:45
andrewzah/openbook
https://api.github.com/repos/andrewzah/openbook
opened
concatenate meter + style metadata headers for lilypond
rust-preprocessor
- [] concatenate meter + style metadata headers
1.0
concatenate meter + style metadata headers for lilypond - - [] concatenate meter + style metadata headers
process
concatenate meter style metadata headers for lilypond concatenate meter style metadata headers
1
62,378
17,023,909,637
IssuesEvent
2021-07-03 04:30:25
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
place:isolated_dwelling refenced as parent for many unrelated objects.
Component: nominatim Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 3.50pm, Tuesday, 26th August 2014]** object http://nominatim.openstreetmap.org/details.php?place_id=9151993414 is being referenced as parent of many completely unrelated objects as you can see from the list on the same page. Don't really know how to correct this problem as there is nothing wrong with the details of the above mentioned object.
1.0
place:isolated_dwelling refenced as parent for many unrelated objects. - **[Submitted to the original trac issue database at 3.50pm, Tuesday, 26th August 2014]** object http://nominatim.openstreetmap.org/details.php?place_id=9151993414 is being referenced as parent of many completely unrelated objects as you can see from the list on the same page. Don't really know how to correct this problem as there is nothing wrong with the details of the above mentioned object.
non_process
place isolated dwelling refenced as parent for many unrelated objects object is being referenced as parent of many completely unrelated objects as you can see from the list on the same page don t really know how to correct this problem as there is nothing wrong with the details of the above mentioned object
0
444,539
31,074,171,010
IssuesEvent
2023-08-12 09:15:47
damies13/rfswarm
https://api.github.com/repos/damies13/rfswarm
opened
Document options for connecting agents to manager
documentation enhancement
**Is your feature request related to a problem? Please describe.** Document options for connecting agents to manager, as it's currently not documented - Port forwarding (NAT?) - dydns (no-ip etc) - ngrok https://ngrok.com/ - Hamachi https://www.vpn.net/ **Describe the solution you'd like** Documentation update **Additional context** Excerpt from discord chat where this came up: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:05]Joomy: Hi all, im currently load testing a site for our project. I would like to ask if it is possible to connect an agent to a manager even if both of them are connected to a different wifi/lan network? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:41]damies13: Hi Joomy, Sure, no problem, that's the intention, all the network traffic between the agents and manager is one way and on 1 port, so even of you have multiple firewalls you only have to open a single port for the manager. The default port on the manager is 8138, but you can change it if you need to ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:43]damies13: as long as the agent can connect to the manager the test will run ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:47]damies13: All the info you need to configure this should be in the documentation, if you're having trouble finding it I can point you in the right direction ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:49]Joomy: Hi! thanks you for answering. May I please know in which part of the documentation includes the configuration? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:57]damies13: Looking back over the documentation, I see it's a bit scattered, in the page on the Manager, the sections you want are https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md#2-adjust-the-firewall I also Noticed there's no documentation on how to adjust the ini file, but if you've ever adjusted an ini file for any other application it should be obvious, the manager's ini file is RFSwarmManager.ini and in there you can change the manager settings if you need to. you can also change it in the manager GUI see https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md#settings-for-the-scenario-and-rfswarm- GitHub [rfswarm/rfswarm_manager.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/rfswarm_manager.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:59]damies13: Once you have the manager setup you can refer to this section for the ini settings on the agent https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_agent.md#ini-file-settings GitHub [rfswarm/rfswarm_agent.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_agent.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/rfswarm_agent.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:00]damies13: Also It's worthwhile reading the Overview section https://github.com/damies13/rfswarm/blob/master/Doc/Overview.md Particularly the section on Robot File handling (transfer from Manager to Agent) GitHub [rfswarm/Overview.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/Overview.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/Overview.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:05]Joomy: Thanks for this. For example i have two wifi connections with different ISPs. I need to configure my firewall on the manager side (connected to wifi-1) and the firewall on the agent side (connected to wifi-2)? Am i correct? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:16]damies13: maybe, it all depends on your network ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:18]damies13: Start with the agent, if the wifi network on the agent side it NAT'ed then there's probably no configuration needed on that side, so that's the easy bit. ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:24]damies13: on the manager side if the manager is on a NAT'ed network, so doesn't have a public IP address, then the easiest thing to do might be port forward the port 8138 on the router to port 8138 on the private IP of the manager, note down the IP address of the router on the manager side, because you'll need to put that in the agent's ini file for the swarmmanager setting. You can use the agent's tab on the manager to verify when the agent has successfully connected (it can take up to 20 seconds for the agent to show up), likewise on the agent you'll see a message like Manager Connected http://localhost:8138/ 2023-03-20 17:36:03 ( 1679297763 ) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:47]Joomy: thank you for this answer. Ill try these out ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:48]Joomy: also is there an alternative way besides port forwarding? I think port forwarding requires to have a static ip ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ >@Joomy also is there an alternative way besides port forwarding? I think port forwarding requires to have a static ip [15:07]damies13: Nothing that doesn't require the IP address to remain the same for the time of the test. You could setup a reverse proxy server, it's basically the same as if you had a web server on you wifi-1 network, how would web browsers on the internet connect to it? I deliberately kept the communication from agent to manager as http traffic to give you the most options. I will mention that the IP on the manager side doesn't have to be static for port forwarding, just remain the same for the test. Many ISP's give you a dhcp address, but the address doesn't actually change for days or weeks, often only changing when your restart your router. So if this is the case for you, then you just need to update the agents ini file when the IP address changes. You could also use a Dynamic DNS Provider (no-ip or similar) and use the host name you register with them in your agents ini file, and let their service keep track of when your ip address changes A another option, if you can't get it to work with IPv4 but both sites have IPv6, I've tested the agent to manager communication on an IPv6 only network and it works that way too, so if you can get an IPv6 address on your manager and the agent can also get an IPv6 address, then you could potentially make it work that way. ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:11]plhrja: One straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic, e.g. ngrok https://ngrok.com/ This however creates extra latency between the agents and the manager, so suboptimal solution performance wise. On the other hand, fast to setup and test your configuration. ngrok - Online in One Line ngrok is the fastest way to put anything on the internet with a single command. 2 ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ >@plhrja One straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic, e.g. ngrok https://ngrok.com/ This however creates extra latency between the agents and the manager, so suboptimal solution performance wise. On the other hand, fast to setup and test your configuration. [15:13]damies13: I've never heard of ngrok, great suggestion ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:22]Joomy: thank you all for this. I actually tried using Hamachi to setup a virtual lan for the two devices and i think its working ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:23]Joomy: i am able to successfully run a test case
1.0
Document options for connecting agents to manager - **Is your feature request related to a problem? Please describe.** Document options for connecting agents to manager, as it's currently not documented - Port forwarding (NAT?) - dydns (no-ip etc) - ngrok https://ngrok.com/ - Hamachi https://www.vpn.net/ **Describe the solution you'd like** Documentation update **Additional context** Excerpt from discord chat where this came up: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:05]Joomy: Hi all, im currently load testing a site for our project. I would like to ask if it is possible to connect an agent to a manager even if both of them are connected to a different wifi/lan network? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:41]damies13: Hi Joomy, Sure, no problem, that's the intention, all the network traffic between the agents and manager is one way and on 1 port, so even of you have multiple firewalls you only have to open a single port for the manager. The default port on the manager is 8138, but you can change it if you need to ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:43]damies13: as long as the agent can connect to the manager the test will run ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:47]damies13: All the info you need to configure this should be in the documentation, if you're having trouble finding it I can point you in the right direction ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:49]Joomy: Hi! thanks you for answering. May I please know in which part of the documentation includes the configuration? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:57]damies13: Looking back over the documentation, I see it's a bit scattered, in the page on the Manager, the sections you want are https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md#2-adjust-the-firewall I also Noticed there's no documentation on how to adjust the ini file, but if you've ever adjusted an ini file for any other application it should be obvious, the manager's ini file is RFSwarmManager.ini and in there you can change the manager settings if you need to. you can also change it in the manager GUI see https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md#settings-for-the-scenario-and-rfswarm- GitHub [rfswarm/rfswarm_manager.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_manager.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/rfswarm_manager.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [13:59]damies13: Once you have the manager setup you can refer to this section for the ini settings on the agent https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_agent.md#ini-file-settings GitHub [rfswarm/rfswarm_agent.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/rfswarm_agent.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/rfswarm_agent.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:00]damies13: Also It's worthwhile reading the Overview section https://github.com/damies13/rfswarm/blob/master/Doc/Overview.md Particularly the section on Robot File handling (transfer from Manager to Agent) GitHub [rfswarm/Overview.md at master · damies13/rfswarm](https://github.com/damies13/rfswarm/blob/master/Doc/Overview.md) Robot Framework Swarm . Contribute to damies13/rfswarm development by creating an account on GitHub. rfswarm/Overview.md at master · damies13/rfswarm ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:05]Joomy: Thanks for this. For example i have two wifi connections with different ISPs. I need to configure my firewall on the manager side (connected to wifi-1) and the firewall on the agent side (connected to wifi-2)? Am i correct? ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:16]damies13: maybe, it all depends on your network ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:18]damies13: Start with the agent, if the wifi network on the agent side it NAT'ed then there's probably no configuration needed on that side, so that's the easy bit. ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:24]damies13: on the manager side if the manager is on a NAT'ed network, so doesn't have a public IP address, then the easiest thing to do might be port forward the port 8138 on the router to port 8138 on the private IP of the manager, note down the IP address of the router on the manager side, because you'll need to put that in the agent's ini file for the swarmmanager setting. You can use the agent's tab on the manager to verify when the agent has successfully connected (it can take up to 20 seconds for the agent to show up), likewise on the agent you'll see a message like Manager Connected http://localhost:8138/ 2023-03-20 17:36:03 ( 1679297763 ) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:47]Joomy: thank you for this answer. Ill try these out ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [14:48]Joomy: also is there an alternative way besides port forwarding? I think port forwarding requires to have a static ip ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ >@Joomy also is there an alternative way besides port forwarding? I think port forwarding requires to have a static ip [15:07]damies13: Nothing that doesn't require the IP address to remain the same for the time of the test. You could setup a reverse proxy server, it's basically the same as if you had a web server on you wifi-1 network, how would web browsers on the internet connect to it? I deliberately kept the communication from agent to manager as http traffic to give you the most options. I will mention that the IP on the manager side doesn't have to be static for port forwarding, just remain the same for the test. Many ISP's give you a dhcp address, but the address doesn't actually change for days or weeks, often only changing when your restart your router. So if this is the case for you, then you just need to update the agents ini file when the IP address changes. You could also use a Dynamic DNS Provider (no-ip or similar) and use the host name you register with them in your agents ini file, and let their service keep track of when your ip address changes A another option, if you can't get it to work with IPv4 but both sites have IPv6, I've tested the agent to manager communication on an IPv6 only network and it works that way too, so if you can get an IPv6 address on your manager and the agent can also get an IPv6 address, then you could potentially make it work that way. ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:11]plhrja: One straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic, e.g. ngrok https://ngrok.com/ This however creates extra latency between the agents and the manager, so suboptimal solution performance wise. On the other hand, fast to setup and test your configuration. ngrok - Online in One Line ngrok is the fastest way to put anything on the internet with a single command. 2 ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ >@plhrja One straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic, e.g. ngrok https://ngrok.com/ This however creates extra latency between the agents and the manager, so suboptimal solution performance wise. On the other hand, fast to setup and test your configuration. [15:13]damies13: I've never heard of ngrok, great suggestion ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:22]Joomy: thank you all for this. I actually tried using Hamachi to setup a virtual lan for the two devices and i think its working ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [15:23]Joomy: i am able to successfully run a test case
non_process
document options for connecting agents to manager is your feature request related to a problem please describe document options for connecting agents to manager as it s currently not documented port forwarding nat dydns no ip etc ngrok hamachi describe the solution you d like documentation update additional context excerpt from discord chat where this came up joomy hi all im currently load testing a site for our project i would like to ask if it is possible to connect an agent to a manager even if both of them are connected to a different wifi lan network hi joomy sure no problem that s the intention all the network traffic between the agents and manager is one way and on port so even of you have multiple firewalls you only have to open a single port for the manager the default port on the manager is but you can change it if you need to as long as the agent can connect to the manager the test will run all the info you need to configure this should be in the documentation if you re having trouble finding it i can point you in the right direction joomy hi thanks you for answering may i please know in which part of the documentation includes the configuration looking back over the documentation i see it s a bit scattered in the page on the manager the sections you want are i also noticed there s no documentation on how to adjust the ini file but if you ve ever adjusted an ini file for any other application it should be obvious the manager s ini file is rfswarmmanager ini and in there you can change the manager settings if you need to you can also change it in the manager gui see github robot framework swarm contribute to rfswarm development by creating an account on github rfswarm rfswarm manager md at master · rfswarm once you have the manager setup you can refer to this section for the ini settings on the agent github robot framework swarm contribute to rfswarm development by creating an account on github rfswarm rfswarm agent md at master · rfswarm also it s worthwhile reading the overview section particularly the section on robot file handling transfer from manager to agent github robot framework swarm contribute to rfswarm development by creating an account on github rfswarm overview md at master · rfswarm joomy thanks for this for example i have two wifi connections with different isps i need to configure my firewall on the manager side connected to wifi and the firewall on the agent side connected to wifi am i correct maybe it all depends on your network start with the agent if the wifi network on the agent side it nat ed then there s probably no configuration needed on that side so that s the easy bit on the manager side if the manager is on a nat ed network so doesn t have a public ip address then the easiest thing to do might be port forward the port on the router to port on the private ip of the manager note down the ip address of the router on the manager side because you ll need to put that in the agent s ini file for the swarmmanager setting you can use the agent s tab on the manager to verify when the agent has successfully connected it can take up to seconds for the agent to show up likewise on the agent you ll see a message like manager connected joomy thank you for this answer ill try these out joomy also is there an alternative way besides port forwarding i think port forwarding requires to have a static ip joomy also is there an alternative way besides port forwarding i think port forwarding requires to have a static ip nothing that doesn t require the ip address to remain the same for the time of the test you could setup a reverse proxy server it s basically the same as if you had a web server on you wifi network how would web browsers on the internet connect to it i deliberately kept the communication from agent to manager as http traffic to give you the most options i will mention that the ip on the manager side doesn t have to be static for port forwarding just remain the same for the test many isp s give you a dhcp address but the address doesn t actually change for days or weeks often only changing when your restart your router so if this is the case for you then you just need to update the agents ini file when the ip address changes you could also use a dynamic dns provider no ip or similar and use the host name you register with them in your agents ini file and let their service keep track of when your ip address changes a another option if you can t get it to work with but both sites have i ve tested the agent to manager communication on an only network and it works that way too so if you can get an address on your manager and the agent can also get an address then you could potentially make it work that way plhrja one straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic e g ngrok this however creates extra latency between the agents and the manager so suboptimal solution performance wise on the other hand fast to setup and test your configuration ngrok online in one line ngrok is the fastest way to put anything on the internet with a single command plhrja one straightforward option for configuring a reverse proxy for the manager is to use a service that provisions a public domain to forward traffic e g ngrok this however creates extra latency between the agents and the manager so suboptimal solution performance wise on the other hand fast to setup and test your configuration i ve never heard of ngrok great suggestion joomy thank you all for this i actually tried using hamachi to setup a virtual lan for the two devices and i think its working joomy i am able to successfully run a test case
0
105,984
4,257,832,783
IssuesEvent
2016-07-11 01:51:31
pdean1/CS6920-Group-4-Project
https://api.github.com/repos/pdean1/CS6920-Group-4-Project
closed
As a user, I'd like to be able to add a budget to the Budget Buddy application so that I can manage that budget
enhancement High Priority
@iledelin
1.0
As a user, I'd like to be able to add a budget to the Budget Buddy application so that I can manage that budget - @iledelin
non_process
as a user i d like to be able to add a budget to the budget buddy application so that i can manage that budget iledelin
0
233,614
7,700,854,473
IssuesEvent
2018-05-20 08:06:37
elysium-project/bug-tracker
https://api.github.com/repos/elysium-project/bug-tracker
opened
Hellfire not triggering Blade of Eternal Darkness
Confirmed High Priority
testing varius AoEs that it does work can someone explain why hellfire shouldn't work?
1.0
Hellfire not triggering Blade of Eternal Darkness - testing varius AoEs that it does work can someone explain why hellfire shouldn't work?
non_process
hellfire not triggering blade of eternal darkness testing varius aoes that it does work can someone explain why hellfire shouldn t work
0
11,758
14,591,886,768
IssuesEvent
2020-12-19 15:06:49
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
TIN interpolation crashes QGIS or creates empty output
Bug Crash/Data Corruption Processing
Add the attached line layer Interpolate the "elevation" attribute using "Break lines" or "structure lines" as type. QGIS 3.10.12 crashes. QGIS 3.16.1 creates an empty output. If using "points" as type works [cn.zip](https://github.com/qgis/QGIS/files/5717883/cn.zip) ok on both versions. Tested on Windows 10.
1.0
TIN interpolation crashes QGIS or creates empty output - Add the attached line layer Interpolate the "elevation" attribute using "Break lines" or "structure lines" as type. QGIS 3.10.12 crashes. QGIS 3.16.1 creates an empty output. If using "points" as type works [cn.zip](https://github.com/qgis/QGIS/files/5717883/cn.zip) ok on both versions. Tested on Windows 10.
process
tin interpolation crashes qgis or creates empty output add the attached line layer interpolate the elevation attribute using break lines or structure lines as type qgis crashes qgis creates an empty output if using points as type works ok on both versions tested on windows
1
609,483
18,874,200,095
IssuesEvent
2021-11-13 18:32:25
BranniganLab/blobulator
https://api.github.com/repos/BranniganLab/blobulator
opened
Miscellaneous Tab Cleanups
top priority
Move "Result" next to "New Query" Add a tab back to the Brannigan Lab Homepage.
1.0
Miscellaneous Tab Cleanups - Move "Result" next to "New Query" Add a tab back to the Brannigan Lab Homepage.
non_process
miscellaneous tab cleanups move result next to new query add a tab back to the brannigan lab homepage
0
13,624
9,999,182,911
IssuesEvent
2019-07-12 09:59:23
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Met internal server err
cognitive-services/svc cxp product-question speech-service/subsvc triaged
Hi, Today when i run the sample code in this page from eastasia region, frequently met below err: StatusCode = InternalServerError ReasonPhrase = Internal Server Error CreateVoiceSignatureByUsingBody() StatusCode = InternalServerError ReasonPhrase = Internal Server Error CreateVoiceSignatureByUsingFormData() Previously running was not smooth either: No any response at all. Today is the first time i saw the response, though w/ error. Can you check it? Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 166f57df-05f7-aacf-3a93-a11e688bd144 * Version Independent ID: c62ef65b-049b-e231-e728-3d2a8f4ddd97 * Content: [Transcribe multi-participant conversations with the Speech SDK - Speech Services - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/how-to-use-conversation-transcription-service) * Content Source: [articles/cognitive-services/Speech-Service/how-to-use-conversation-transcription-service.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Speech-Service/how-to-use-conversation-transcription-service.md) * Service: **cognitive-services** * Sub-service: **speech-service** * GitHub Login: @jhakulin * Microsoft Alias: **jhakulin**
2.0
Met internal server err - Hi, Today when i run the sample code in this page from eastasia region, frequently met below err: StatusCode = InternalServerError ReasonPhrase = Internal Server Error CreateVoiceSignatureByUsingBody() StatusCode = InternalServerError ReasonPhrase = Internal Server Error CreateVoiceSignatureByUsingFormData() Previously running was not smooth either: No any response at all. Today is the first time i saw the response, though w/ error. Can you check it? Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 166f57df-05f7-aacf-3a93-a11e688bd144 * Version Independent ID: c62ef65b-049b-e231-e728-3d2a8f4ddd97 * Content: [Transcribe multi-participant conversations with the Speech SDK - Speech Services - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/how-to-use-conversation-transcription-service) * Content Source: [articles/cognitive-services/Speech-Service/how-to-use-conversation-transcription-service.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Speech-Service/how-to-use-conversation-transcription-service.md) * Service: **cognitive-services** * Sub-service: **speech-service** * GitHub Login: @jhakulin * Microsoft Alias: **jhakulin**
non_process
met internal server err hi today when i run the sample code in this page from eastasia region frequently met below err statuscode internalservererror reasonphrase internal server error createvoicesignaturebyusingbody statuscode internalservererror reasonphrase internal server error createvoicesignaturebyusingformdata previously running was not smooth either no any response at all today is the first time i saw the response though w error can you check it thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id aacf version independent id content content source service cognitive services sub service speech service github login jhakulin microsoft alias jhakulin
0
120,489
17,644,204,402
IssuesEvent
2021-08-20 01:56:56
m0lese/Linux-Phaco
https://api.github.com/repos/m0lese/Linux-Phaco
opened
WS-2021-0315 (High) detected in linuxv5.12.16
security vulnerability
## WS-2021-0315 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.12.16</b></p></summary> <p> <p>Linux kernel stable tree mirror</p> <p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p> <p>Found in base branch: <b>prime</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>Linux-Phaco/drivers/scsi/libiscsi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Linux/Kernel in versions v5.13-rc1 to v5.13.3 is vulnerable to conn use after free during resets <p>Publish Date: 2021-05-31 <p>URL: <a href=https://github.com/gregkh/linux/commit/fa9542b35ceb4202e8f8d65f440529a63524dca9>WS-2021-0315</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001200">https://osv.dev/vulnerability/UVI-2021-1001200</a></p> <p>Release Date: 2021-05-31</p> <p>Fix Resolution: v5.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0315 (High) detected in linuxv5.12.16 - ## WS-2021-0315 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.12.16</b></p></summary> <p> <p>Linux kernel stable tree mirror</p> <p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p> <p>Found in base branch: <b>prime</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>Linux-Phaco/drivers/scsi/libiscsi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Linux/Kernel in versions v5.13-rc1 to v5.13.3 is vulnerable to conn use after free during resets <p>Publish Date: 2021-05-31 <p>URL: <a href=https://github.com/gregkh/linux/commit/fa9542b35ceb4202e8f8d65f440529a63524dca9>WS-2021-0315</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001200">https://osv.dev/vulnerability/UVI-2021-1001200</a></p> <p>Release Date: 2021-05-31</p> <p>Fix Resolution: v5.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in ws high severity vulnerability vulnerable library linux kernel stable tree mirror library home page a href found in base branch prime vulnerable source files linux phaco drivers scsi libiscsi c vulnerability details linux kernel in versions to is vulnerable to conn use after free during resets publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
532,639
15,560,566,750
IssuesEvent
2021-03-16 12:52:09
canonical-web-and-design/vanilla-framework
https://api.github.com/repos/canonical-web-and-design/vanilla-framework
reopened
Standards for table patterns
Epic Priority: Medium WG: Validated
Bartek created a new pattern for snapcraft-flask, [`p-table-key-value`](https://github.com/canonical-websites/snapcraft-flask/pull/51), which follows precedent from existing table patterns in Vanilla pretty well. However, I'm not too happy about this precedent: - there is one actual pattern, `p-table-expanding`, and two modifiers for "`p-table`" - `p-table--sortable` and `p-table--mobile-card` - all three patterns actually target `td` and `tr` elements directly I thought it was policy not to target elements directly. And so these should ideally be targeting e.g. `.p-table-expanding-row` and `.p-table-expanding-cell` instead of `tr` and `td`. I think these patterns may be out-of-date because they were imported from themes with less strict standards, but I just wanted to confirm if my understanding of the ideal implementation is correct or not. I don't think modifiers based on a `p-table` pattern should exist if the `p-table` pattern does not itself exist. And so e.g. `p-table--sortable` should really be `p-table-sortable`. This brings up the issue that `p-table-expanding` in fact looks incredibly similar to `p-table--sortable`, even though the former is a pattern rather than a modifier. One could be forgiven for confusing patterns named like this with modifiers. So I'd also like to suggest that when we create *patterns* with adjectives in the name, we consider using the more natural form of `p-{adjective}-{noun}` rather than `p-{noun}-{adjective}`, to clearly distinguish them from modifiers. Then `p-expanding-table` would be clearly of a different type than `p-table--sortable`. Thoughts?
1.0
Standards for table patterns - Bartek created a new pattern for snapcraft-flask, [`p-table-key-value`](https://github.com/canonical-websites/snapcraft-flask/pull/51), which follows precedent from existing table patterns in Vanilla pretty well. However, I'm not too happy about this precedent: - there is one actual pattern, `p-table-expanding`, and two modifiers for "`p-table`" - `p-table--sortable` and `p-table--mobile-card` - all three patterns actually target `td` and `tr` elements directly I thought it was policy not to target elements directly. And so these should ideally be targeting e.g. `.p-table-expanding-row` and `.p-table-expanding-cell` instead of `tr` and `td`. I think these patterns may be out-of-date because they were imported from themes with less strict standards, but I just wanted to confirm if my understanding of the ideal implementation is correct or not. I don't think modifiers based on a `p-table` pattern should exist if the `p-table` pattern does not itself exist. And so e.g. `p-table--sortable` should really be `p-table-sortable`. This brings up the issue that `p-table-expanding` in fact looks incredibly similar to `p-table--sortable`, even though the former is a pattern rather than a modifier. One could be forgiven for confusing patterns named like this with modifiers. So I'd also like to suggest that when we create *patterns* with adjectives in the name, we consider using the more natural form of `p-{adjective}-{noun}` rather than `p-{noun}-{adjective}`, to clearly distinguish them from modifiers. Then `p-expanding-table` would be clearly of a different type than `p-table--sortable`. Thoughts?
non_process
standards for table patterns bartek created a new pattern for snapcraft flask which follows precedent from existing table patterns in vanilla pretty well however i m not too happy about this precedent there is one actual pattern p table expanding and two modifiers for p table p table sortable and p table mobile card all three patterns actually target td and tr elements directly i thought it was policy not to target elements directly and so these should ideally be targeting e g p table expanding row and p table expanding cell instead of tr and td i think these patterns may be out of date because they were imported from themes with less strict standards but i just wanted to confirm if my understanding of the ideal implementation is correct or not i don t think modifiers based on a p table pattern should exist if the p table pattern does not itself exist and so e g p table sortable should really be p table sortable this brings up the issue that p table expanding in fact looks incredibly similar to p table sortable even though the former is a pattern rather than a modifier one could be forgiven for confusing patterns named like this with modifiers so i d also like to suggest that when we create patterns with adjectives in the name we consider using the more natural form of p adjective noun rather than p noun adjective to clearly distinguish them from modifiers then p expanding table would be clearly of a different type than p table sortable thoughts
0
6,784
9,917,115,663
IssuesEvent
2019-06-28 22:29:59
HumanCellAtlas/dcp-community
https://api.github.com/repos/HumanCellAtlas/dcp-community
closed
Improving community notifications for RFCs
rfc-process
- [x] Delete references to _This week in DCP_ - [ ] Use #dcp-announce[ments] instead of #dcp channel for rfc _state changes_
1.0
Improving community notifications for RFCs - - [x] Delete references to _This week in DCP_ - [ ] Use #dcp-announce[ments] instead of #dcp channel for rfc _state changes_
process
improving community notifications for rfcs delete references to this week in dcp use dcp announce instead of dcp channel for rfc state changes
1
98,820
20,799,568,905
IssuesEvent
2022-03-17 12:42:20
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
closed
Short solution needed: "Authentication with Nginx and htpasswd" (nginx)
help wanted good first issue code nginx
Please help us write most modern and shortest code solution for this issue: **Authentication with Nginx and htpasswd** (technology: [nginx](https://onelinerhub.com/nginx)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "Authentication with Nginx and htpasswd" (nginx) - Please help us write most modern and shortest code solution for this issue: **Authentication with Nginx and htpasswd** (technology: [nginx](https://onelinerhub.com/nginx)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
non_process
short solution needed authentication with nginx and htpasswd nginx please help us write most modern and shortest code solution for this issue authentication with nginx and htpasswd technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request
0
12,812
15,189,070,366
IssuesEvent
2021-02-15 15:55:22
Geonovum/disgeo-arch
https://api.github.com/repos/Geonovum/disgeo-arch
reopened
Maak Validatie zichtbaar
In Behandeling In behandeling - voorstel processen e.d. Processen Functies Componenten
Wellicht kan de functie Validatie separaat van de functie Registratie in de hoofdstructuur zichtbaar gemaakt worden. Voor alle partijen moet het mogelijk zijn om te valideren of gegevens voldoen aan de gestelde vereisten, zowel voorafgaand aan registratie als tijdens registratie als bij of na afname van gegevens.
2.0
Maak Validatie zichtbaar - Wellicht kan de functie Validatie separaat van de functie Registratie in de hoofdstructuur zichtbaar gemaakt worden. Voor alle partijen moet het mogelijk zijn om te valideren of gegevens voldoen aan de gestelde vereisten, zowel voorafgaand aan registratie als tijdens registratie als bij of na afname van gegevens.
process
maak validatie zichtbaar wellicht kan de functie validatie separaat van de functie registratie in de hoofdstructuur zichtbaar gemaakt worden voor alle partijen moet het mogelijk zijn om te valideren of gegevens voldoen aan de gestelde vereisten zowel voorafgaand aan registratie als tijdens registratie als bij of na afname van gegevens
1
20,140
28,141,322,174
IssuesEvent
2023-04-02 00:43:05
4drian3d/KickRedirect
https://api.github.com/repos/4drian3d/KickRedirect
closed
Use a higher priority on KickedFromServerEvent
good first issue incompatibility
Currently, the plugin uses the `EARLY` priority level. This can cause conflicts with utility plugins that implement a similar feature but intentionally use `EARLY` so they can be overridden. Considering that this plugin has no other use than redirecting players after kick, it would be IMO better to use `NORMAL` or even `LATE`.
True
Use a higher priority on KickedFromServerEvent - Currently, the plugin uses the `EARLY` priority level. This can cause conflicts with utility plugins that implement a similar feature but intentionally use `EARLY` so they can be overridden. Considering that this plugin has no other use than redirecting players after kick, it would be IMO better to use `NORMAL` or even `LATE`.
non_process
use a higher priority on kickedfromserverevent currently the plugin uses the early priority level this can cause conflicts with utility plugins that implement a similar feature but intentionally use early so they can be overridden considering that this plugin has no other use than redirecting players after kick it would be imo better to use normal or even late
0
321,620
23,864,055,130
IssuesEvent
2022-09-07 09:30:00
sunpy/sunpy-soar
https://api.github.com/repos/sunpy/sunpy-soar
closed
What data products are already duplicated with the VSO?
documentation
### Provide a general description of the issue or problem. The VSO is already indexing some of the SOAR archive, this issue is to keep track of what products are and are not also downloadable through the VSO. From speaking to Ed, I know that the EUI data are already accessible (without the ability to search on Level), if anyone knows anything else please let us know.
1.0
What data products are already duplicated with the VSO? - ### Provide a general description of the issue or problem. The VSO is already indexing some of the SOAR archive, this issue is to keep track of what products are and are not also downloadable through the VSO. From speaking to Ed, I know that the EUI data are already accessible (without the ability to search on Level), if anyone knows anything else please let us know.
non_process
what data products are already duplicated with the vso provide a general description of the issue or problem the vso is already indexing some of the soar archive this issue is to keep track of what products are and are not also downloadable through the vso from speaking to ed i know that the eui data are already accessible without the ability to search on level if anyone knows anything else please let us know
0
6,815
3,909,010,749
IssuesEvent
2016-04-19 17:46:25
uProxy/uproxy
https://api.github.com/repos/uProxy/uproxy
opened
third_party handling is a mess
C:BuildProcess P2
Problem is uproxy's `typings.json` has to be a **superset** of uproxy, owing to how uproxy's `setup.sh` "merges" the two repos' folders. Several ways to fix this, not sure which is best: - `typings` folder should be inside `src` - move uproxy-lib into uproxy and forget about this
1.0
third_party handling is a mess - Problem is uproxy's `typings.json` has to be a **superset** of uproxy, owing to how uproxy's `setup.sh` "merges" the two repos' folders. Several ways to fix this, not sure which is best: - `typings` folder should be inside `src` - move uproxy-lib into uproxy and forget about this
non_process
third party handling is a mess problem is uproxy s typings json has to be a superset of uproxy owing to how uproxy s setup sh merges the two repos folders several ways to fix this not sure which is best typings folder should be inside src move uproxy lib into uproxy and forget about this
0
5,267
3,917,001,967
IssuesEvent
2016-04-21 05:57:13
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_Neighbourhood Operations_Back to experiment_smk
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description : In the "Neighbourhood Operations" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user Actual Result : In the "Neighbourhood Operations" experiment,the back to experiments link is not present in the page Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/image-processing-iiith/blob/master/test-cases/integration_test-cases/Neighbourhood%20Operations/Neighbourhood%20Operations_20_Back%20to%20experiment_smk.org
True
QA_Neighbourhood Operations_Back to experiment_smk - Defect Description : In the "Neighbourhood Operations" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user Actual Result : In the "Neighbourhood Operations" experiment,the back to experiments link is not present in the page Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/image-processing-iiith/blob/master/test-cases/integration_test-cases/Neighbourhood%20Operations/Neighbourhood%20Operations_20_Back%20to%20experiment_smk.org
non_process
qa neighbourhood operations back to experiment smk defect description in the neighbourhood operations experiment the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in order to view the list of experiments by the user actual result in the neighbourhood operations experiment the back to experiments link is not present in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
0
123,850
17,772,349,517
IssuesEvent
2021-08-30 14:59:46
kapseliboi/spotlight
https://api.github.com/repos/kapseliboi/spotlight
opened
CVE-2018-11697 (High) detected in multiple libraries
security vulnerability
## CVE-2018-11697 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>guibalaguibala</b>, <b>CSS::Sass3.6.3</b>, <b>CSS::Sass3.6.3</b>, <b>node-sass-4.5.3.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.5.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.5.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.5.3.tgz</a></p> <p>Path to dependency file: spotlight/package.json</p> <p>Path to vulnerable library: spotlight/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - grunt-sass-2.0.0.tgz (Root Library) - :x: **node-sass-4.5.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/spotlight/commit/efa8ebd4395408150b8ea1a18eec77751d13827b">efa8ebd4395408150b8ea1a18eec77751d13827b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::Prelexer::exactly() which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11697>CVE-2018-11697</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11697 (High) detected in multiple libraries - ## CVE-2018-11697 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>guibalaguibala</b>, <b>CSS::Sass3.6.3</b>, <b>CSS::Sass3.6.3</b>, <b>node-sass-4.5.3.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.5.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.5.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.5.3.tgz</a></p> <p>Path to dependency file: spotlight/package.json</p> <p>Path to vulnerable library: spotlight/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - grunt-sass-2.0.0.tgz (Root Library) - :x: **node-sass-4.5.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/spotlight/commit/efa8ebd4395408150b8ea1a18eec77751d13827b">efa8ebd4395408150b8ea1a18eec77751d13827b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::Prelexer::exactly() which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11697>CVE-2018-11697</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries guibalaguibala css css node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file spotlight package json path to vulnerable library spotlight node modules node sass package json dependency hierarchy grunt sass tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass prelexer exactly which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
17,392
23,208,277,879
IssuesEvent
2022-08-02 07:53:38
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] In batch mode, allow population of file/layer input columns by searching for files matching a specified pattern
Automatic new feature Processing 3.8
Original commit: https://github.com/qgis/QGIS/commit/e6e39173b2b94fb7387b0c8bb1748fd87e4f5e05 by nyalldawson With optional recursive search!
1.0
[FEATURE][processing] In batch mode, allow population of file/layer input columns by searching for files matching a specified pattern - Original commit: https://github.com/qgis/QGIS/commit/e6e39173b2b94fb7387b0c8bb1748fd87e4f5e05 by nyalldawson With optional recursive search!
process
in batch mode allow population of file layer input columns by searching for files matching a specified pattern original commit by nyalldawson with optional recursive search
1
241,131
7,809,030,326
IssuesEvent
2018-06-11 22:20:14
eustasy/bubbly
https://api.github.com/repos/eustasy/bubbly
opened
Speed up DH Parameter generation.
Priority: High Status: Confirmed Type: Enhancement
### Option 1: Use the `dsaparam` flag on generation. >The reasonable solution would be to add the -dsaparam option. > `openssl dhparam -dsaparam -out /etc/ssl/private/dhparam.pem 4096` > This option instructs OpenSSL to produce "DSA-like" DH parameters (p is such that p-1 is a multiple of a smaller prime q, and the generator has multiplicative order q). This is considerably faster because it does not need to nest the primality tests, and thus only thousands, not millions, of candidates will be generated and tested. > As far as academics know, DSA-like parameters for DH are equally secure; there is no actual advantage to using "strong primes" (the terminology is traditional and does not actually imply some extra strength). > Similarly, you may also use a 2048-bit modulus, which is already very far into the "cannot break it zone". The 4096-bit modulus will make DH computations slower (which is not a real problem for a VPN; these occur only at the start of the connection), but won't actually improve security. > To some extent, a 4096-bit modulus may woo auditors, but auditors are unlikely to be much impressed by a Raspberry-Pi, which is way too cheap anyway. [Source](https://security.stackexchange.com/a/95184/19792) ### Option 2. Use a service `curl https://2ton.com.au/dhparam/4096` ### Option 3. Install a randomness generator like `rng-tools` See https://www.cyberciti.biz/open-source/debian-ubuntu-centos-linux-setup-additional-entropy-for-server-using-aveged-rng-tools-utils/
1.0
Speed up DH Parameter generation. - ### Option 1: Use the `dsaparam` flag on generation. >The reasonable solution would be to add the -dsaparam option. > `openssl dhparam -dsaparam -out /etc/ssl/private/dhparam.pem 4096` > This option instructs OpenSSL to produce "DSA-like" DH parameters (p is such that p-1 is a multiple of a smaller prime q, and the generator has multiplicative order q). This is considerably faster because it does not need to nest the primality tests, and thus only thousands, not millions, of candidates will be generated and tested. > As far as academics know, DSA-like parameters for DH are equally secure; there is no actual advantage to using "strong primes" (the terminology is traditional and does not actually imply some extra strength). > Similarly, you may also use a 2048-bit modulus, which is already very far into the "cannot break it zone". The 4096-bit modulus will make DH computations slower (which is not a real problem for a VPN; these occur only at the start of the connection), but won't actually improve security. > To some extent, a 4096-bit modulus may woo auditors, but auditors are unlikely to be much impressed by a Raspberry-Pi, which is way too cheap anyway. [Source](https://security.stackexchange.com/a/95184/19792) ### Option 2. Use a service `curl https://2ton.com.au/dhparam/4096` ### Option 3. Install a randomness generator like `rng-tools` See https://www.cyberciti.biz/open-source/debian-ubuntu-centos-linux-setup-additional-entropy-for-server-using-aveged-rng-tools-utils/
non_process
speed up dh parameter generation option use the dsaparam flag on generation the reasonable solution would be to add the dsaparam option openssl dhparam dsaparam out etc ssl private dhparam pem this option instructs openssl to produce dsa like dh parameters p is such that p is a multiple of a smaller prime q and the generator has multiplicative order q this is considerably faster because it does not need to nest the primality tests and thus only thousands not millions of candidates will be generated and tested as far as academics know dsa like parameters for dh are equally secure there is no actual advantage to using strong primes the terminology is traditional and does not actually imply some extra strength similarly you may also use a bit modulus which is already very far into the cannot break it zone the bit modulus will make dh computations slower which is not a real problem for a vpn these occur only at the start of the connection but won t actually improve security to some extent a bit modulus may woo auditors but auditors are unlikely to be much impressed by a raspberry pi which is way too cheap anyway option use a service curl option install a randomness generator like rng tools see
0
123,461
12,198,775,527
IssuesEvent
2020-04-29 23:44:12
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
opened
Logarithmic Axes does not support a ticks.min value of 0 (Zero)
type: documentation
Documentation Is: <!-- Please place an x (no spaces!) in all [ ] that apply --> - [x] Missing or needed - [ ] Confusing - [ ] Not Sure? ### Please Explain in Detail... I needed to produce a number of charts with a consistent yAxes logarithmic scale so that users can compare between charts. I tried setting the same yAxes config settings for each chart: beginAtZero = true ticks.min = 0 ticks.max = <maximum-possible-value> However the scale varied between charts depending on the data values. See example. After much searching and trial and error it appears that the ticks.min=0 setting is ignored (probably because the logarithm of zero is not defined). ### Your Proposal for Changes In tick Configuration (https://www.chartjs.org/docs/latest/axes/cartesian/#tick-configuration) for 'min' description add a note such as: (Note: logarithmic scales do not support minimum value of zero) ### Example https://codepen.io/phil-howell/pen/KKdvoKW
1.0
Logarithmic Axes does not support a ticks.min value of 0 (Zero) - Documentation Is: <!-- Please place an x (no spaces!) in all [ ] that apply --> - [x] Missing or needed - [ ] Confusing - [ ] Not Sure? ### Please Explain in Detail... I needed to produce a number of charts with a consistent yAxes logarithmic scale so that users can compare between charts. I tried setting the same yAxes config settings for each chart: beginAtZero = true ticks.min = 0 ticks.max = <maximum-possible-value> However the scale varied between charts depending on the data values. See example. After much searching and trial and error it appears that the ticks.min=0 setting is ignored (probably because the logarithm of zero is not defined). ### Your Proposal for Changes In tick Configuration (https://www.chartjs.org/docs/latest/axes/cartesian/#tick-configuration) for 'min' description add a note such as: (Note: logarithmic scales do not support minimum value of zero) ### Example https://codepen.io/phil-howell/pen/KKdvoKW
non_process
logarithmic axes does not support a ticks min value of zero documentation is missing or needed confusing not sure please explain in detail i needed to produce a number of charts with a consistent yaxes logarithmic scale so that users can compare between charts i tried setting the same yaxes config settings for each chart beginatzero true ticks min ticks max however the scale varied between charts depending on the data values see example after much searching and trial and error it appears that the ticks min setting is ignored probably because the logarithm of zero is not defined your proposal for changes in tick configuration for min description add a note such as note logarithmic scales do not support minimum value of zero example
0
284,658
8,745,063,305
IssuesEvent
2018-12-13 00:51:59
netdata/netdata
https://api.github.com/repos/netdata/netdata
closed
[bug]some metrics don't report to /allmetrics endpoint with prometheus format
area/backends bug priority/medium
# What I did 1. installed memcached and start it 1. installed netdata and start it # What I see * memcached charts showed on dashboard * memcached metrics showed on `/api/v1/allmetrics?format=shell` ```bash # chart: memcached_11211.delete (name: memcached_11211.delete) NETDATA_MEMCACHED_11211_DELETE_HITS="0" # requests NETDATA_MEMCACHED_11211_DELETE_MISSES="0" # requests NETDATA_MEMCACHED_11211_DELETE_VISIBLETOTAL="0" # requests ``` * memcached metrics didn't show on `/api/v1/allmetrics?format=prometheus&help=yes` * only comment, no metrics * `help=yes` or absent don't affect the results ```bash # COMMENT homogeneus chart "memcached_11211.delete", context "memcached.delete", family "delete ops", units "requests" # COMMENT homogeneus chart "memcached_11211.set_rate", context "memcached.set_rate", family "set ops", units "requests/s" # COMMENT homogeneus chart "memcached_11211.net", context "memcached.net", family "network", units "kilobits/s" ``` * After I restart the netdata, all goes well. *I have 2 servers have the same problem. I restarted one and recovered and left the other as is to debug.*
1.0
[bug]some metrics don't report to /allmetrics endpoint with prometheus format - # What I did 1. installed memcached and start it 1. installed netdata and start it # What I see * memcached charts showed on dashboard * memcached metrics showed on `/api/v1/allmetrics?format=shell` ```bash # chart: memcached_11211.delete (name: memcached_11211.delete) NETDATA_MEMCACHED_11211_DELETE_HITS="0" # requests NETDATA_MEMCACHED_11211_DELETE_MISSES="0" # requests NETDATA_MEMCACHED_11211_DELETE_VISIBLETOTAL="0" # requests ``` * memcached metrics didn't show on `/api/v1/allmetrics?format=prometheus&help=yes` * only comment, no metrics * `help=yes` or absent don't affect the results ```bash # COMMENT homogeneus chart "memcached_11211.delete", context "memcached.delete", family "delete ops", units "requests" # COMMENT homogeneus chart "memcached_11211.set_rate", context "memcached.set_rate", family "set ops", units "requests/s" # COMMENT homogeneus chart "memcached_11211.net", context "memcached.net", family "network", units "kilobits/s" ``` * After I restart the netdata, all goes well. *I have 2 servers have the same problem. I restarted one and recovered and left the other as is to debug.*
non_process
some metrics don t report to allmetrics endpoint with prometheus format what i did installed memcached and start it installed netdata and start it what i see memcached charts showed on dashboard memcached metrics showed on api allmetrics format shell bash chart memcached delete name memcached delete netdata memcached delete hits requests netdata memcached delete misses requests netdata memcached delete visibletotal requests memcached metrics didn t show on api allmetrics format prometheus help yes only comment no metrics help yes or absent don t affect the results bash comment homogeneus chart memcached delete context memcached delete family delete ops units requests comment homogeneus chart memcached set rate context memcached set rate family set ops units requests s comment homogeneus chart memcached net context memcached net family network units kilobits s after i restart the netdata all goes well i have servers have the same problem i restarted one and recovered and left the other as is to debug
0
70,884
8,589,649,916
IssuesEvent
2018-11-14 16:56:04
LiskHQ/lisk-mobile
https://api.github.com/repos/LiskHQ/lisk-mobile
closed
Implement the dark mode of the home screen
*hard design enhancement
This includes the implementation of the dark mode for: - Header. - Tab bar. - Account summary. - transaction items.
1.0
Implement the dark mode of the home screen - This includes the implementation of the dark mode for: - Header. - Tab bar. - Account summary. - transaction items.
non_process
implement the dark mode of the home screen this includes the implementation of the dark mode for header tab bar account summary transaction items
0
1,928
4,761,371,732
IssuesEvent
2016-10-25 08:01:08
paulkornikov/Pragonas
https://api.github.com/repos/paulkornikov/Pragonas
closed
Automatiser la prolongation des budgets
a-new feature budget contrat processus workload II
à la connexion et sur check de date... si prolongation autorisée et possible: - récupérer les budgets à terme - pour chaque budget, service de prolongation d'un budget - tracer le processus
1.0
Automatiser la prolongation des budgets - à la connexion et sur check de date... si prolongation autorisée et possible: - récupérer les budgets à terme - pour chaque budget, service de prolongation d'un budget - tracer le processus
process
automatiser la prolongation des budgets à la connexion et sur check de date si prolongation autorisée et possible récupérer les budgets à terme pour chaque budget service de prolongation d un budget tracer le processus
1
13,247
15,715,675,443
IssuesEvent
2021-03-28 02:43:15
Andon-A/New-Hoard-Generator
https://api.github.com/repos/Andon-A/New-Hoard-Generator
closed
Single and Plural Descriptions
In Process enhancement
Currently, each different item has a description, but in many cases these are copied from one to another depending on if the base item is "singular" (A shield) or plural (A pair of bracers). Having a simple system of "singular" and "plural" descriptions could simplify things a lot. Individual descriptions could override them as needed, which would be significantly less often. This would also alleviate the issue of ammunition having grammatically weird descriptions if they have their quantity reduced to 1.
1.0
Single and Plural Descriptions - Currently, each different item has a description, but in many cases these are copied from one to another depending on if the base item is "singular" (A shield) or plural (A pair of bracers). Having a simple system of "singular" and "plural" descriptions could simplify things a lot. Individual descriptions could override them as needed, which would be significantly less often. This would also alleviate the issue of ammunition having grammatically weird descriptions if they have their quantity reduced to 1.
process
single and plural descriptions currently each different item has a description but in many cases these are copied from one to another depending on if the base item is singular a shield or plural a pair of bracers having a simple system of singular and plural descriptions could simplify things a lot individual descriptions could override them as needed which would be significantly less often this would also alleviate the issue of ammunition having grammatically weird descriptions if they have their quantity reduced to
1
133,545
12,543,863,846
IssuesEvent
2020-06-05 16:15:33
Azure/azure-cosmos-dotnet-v3
https://api.github.com/repos/Azure/azure-cosmos-dotnet-v3
closed
GetItemQueryStreamIterator API Name, Casing, and Sample Issues
VNext documentation
Following the docs here: https://docs.microsoft.com/en-us/dotnet/api/azure.cosmos.cosmoscontainer.getitemquerystreamiterator?view=azure-dotnet-preview I was expecting this code: ```csharp public class ToDoActivity{ public string id {get; set;} public string status {get; set;} public int cost {get; set;} } QueryDefinition queryDefinition = new QueryDefinition("select * from ToDos t where t.cost > @expensive") .WithParameter("@expensive", 9000); await foreach(Response response in this.Container.GetItemQueryStreamIterator( queryDefinition, null, new QueryRequestOptions() { PartitionKey = new PartitionKey("Error")})) { using (StreamReader sr = new StreamReader(response.Content)) using (JsonTextReader jtr = new JsonTextReader(sr)) { JObject result = JObject.Load(jtr); } } ``` To return each Document inside of the await foreach loop. But, it actually returns a structure like this: ``` { "Documents": [] } ``` So, I had to create a class like this: ```csharp public class QueryStream { [JsonPropertyName("Documents")] public QueueMessage[] Documents { get; set; } } ``` And then deserialize using that class. ## Issue # 1: "Documents" casing Because "Documents" is Pascal casing it will conflict with my SerializationOptions and child documents, which have camelCasing. > Question: Should we use "documents" instead or enable the user to specify casing for the outer documents. Or ask users to name their property "public QueueMessage[] documents" so the casing is honor. As you can see above I needed to add a JsonPropertyNames attribute, because I'm deserializing with camel and so it expects Documents to be camel as well. ## Issue # 2: API name Based on the example and the API name, I was expecting it to yield for every document, not every page. > Question: Should this be renamed to GetItemPageQueryStreamIterator? ## Issue # 3: Docs are misleading. Because the doc has a single item class ToDoActivity, I was led to believe it would return individual items. > Suggestion: Update the sample to make it very clear what the dev has to do to use it. Here's the code to get this working...this included a lot of trial and error and debugging to figure out. ```csharp using System; using System.Collections.Generic; using System.IO; using System.Text.Json; using System.Text.Json.Serialization; using System.Threading.Tasks; using Azure; using Azure.Cosmos; using Azure.Cosmos.Serialization; using DotNetEnv; namespace cosmostest { public class QueryStream { [JsonPropertyName("Documents")] public QueueMessage[] Documents { get; set; } } class Program { static async Task Main(string[] args) { Env.Load(); CosmosClientOptions options = new CosmosClientOptions { SerializerOptions = new CosmosSerializationOptions { PropertyNamingPolicy = CosmosPropertyNamingPolicy.Default } }; CosmosClient cosmosClient = new CosmosClient( Environment.GetEnvironmentVariable("AZURE_COSMOS_ENDPOINT"), Environment.GetEnvironmentVariable("AZURE_COSMOS_KEY"), options); CosmosContainer cosmosContainer = cosmosClient.GetDatabase(Environment.GetEnvironmentVariable("AZURE_COSMOS_DB")).GetContainer(Environment.GetEnvironmentVariable("AZURE_COSMOS_CONTAINER")); QueryDefinition queryDefinition = new QueryDefinition("SELECT * FROM c"); List<QueueMessage> msgs = new List<QueueMessage>(); await foreach (Response response in cosmosContainer.GetItemQueryStreamIterator(queryDefinition)) { var queryStream = await JsonSerializer.DeserializeAsync<QueryStream>(response.ContentStream, new JsonSerializerOptions() { PropertyNamingPolicy = JsonNamingPolicy.CamelCase }); msgs.AddRange(queryStream.Documents); } } } } ```
1.0
GetItemQueryStreamIterator API Name, Casing, and Sample Issues - Following the docs here: https://docs.microsoft.com/en-us/dotnet/api/azure.cosmos.cosmoscontainer.getitemquerystreamiterator?view=azure-dotnet-preview I was expecting this code: ```csharp public class ToDoActivity{ public string id {get; set;} public string status {get; set;} public int cost {get; set;} } QueryDefinition queryDefinition = new QueryDefinition("select * from ToDos t where t.cost > @expensive") .WithParameter("@expensive", 9000); await foreach(Response response in this.Container.GetItemQueryStreamIterator( queryDefinition, null, new QueryRequestOptions() { PartitionKey = new PartitionKey("Error")})) { using (StreamReader sr = new StreamReader(response.Content)) using (JsonTextReader jtr = new JsonTextReader(sr)) { JObject result = JObject.Load(jtr); } } ``` To return each Document inside of the await foreach loop. But, it actually returns a structure like this: ``` { "Documents": [] } ``` So, I had to create a class like this: ```csharp public class QueryStream { [JsonPropertyName("Documents")] public QueueMessage[] Documents { get; set; } } ``` And then deserialize using that class. ## Issue # 1: "Documents" casing Because "Documents" is Pascal casing it will conflict with my SerializationOptions and child documents, which have camelCasing. > Question: Should we use "documents" instead or enable the user to specify casing for the outer documents. Or ask users to name their property "public QueueMessage[] documents" so the casing is honor. As you can see above I needed to add a JsonPropertyNames attribute, because I'm deserializing with camel and so it expects Documents to be camel as well. ## Issue # 2: API name Based on the example and the API name, I was expecting it to yield for every document, not every page. > Question: Should this be renamed to GetItemPageQueryStreamIterator? ## Issue # 3: Docs are misleading. Because the doc has a single item class ToDoActivity, I was led to believe it would return individual items. > Suggestion: Update the sample to make it very clear what the dev has to do to use it. Here's the code to get this working...this included a lot of trial and error and debugging to figure out. ```csharp using System; using System.Collections.Generic; using System.IO; using System.Text.Json; using System.Text.Json.Serialization; using System.Threading.Tasks; using Azure; using Azure.Cosmos; using Azure.Cosmos.Serialization; using DotNetEnv; namespace cosmostest { public class QueryStream { [JsonPropertyName("Documents")] public QueueMessage[] Documents { get; set; } } class Program { static async Task Main(string[] args) { Env.Load(); CosmosClientOptions options = new CosmosClientOptions { SerializerOptions = new CosmosSerializationOptions { PropertyNamingPolicy = CosmosPropertyNamingPolicy.Default } }; CosmosClient cosmosClient = new CosmosClient( Environment.GetEnvironmentVariable("AZURE_COSMOS_ENDPOINT"), Environment.GetEnvironmentVariable("AZURE_COSMOS_KEY"), options); CosmosContainer cosmosContainer = cosmosClient.GetDatabase(Environment.GetEnvironmentVariable("AZURE_COSMOS_DB")).GetContainer(Environment.GetEnvironmentVariable("AZURE_COSMOS_CONTAINER")); QueryDefinition queryDefinition = new QueryDefinition("SELECT * FROM c"); List<QueueMessage> msgs = new List<QueueMessage>(); await foreach (Response response in cosmosContainer.GetItemQueryStreamIterator(queryDefinition)) { var queryStream = await JsonSerializer.DeserializeAsync<QueryStream>(response.ContentStream, new JsonSerializerOptions() { PropertyNamingPolicy = JsonNamingPolicy.CamelCase }); msgs.AddRange(queryStream.Documents); } } } } ```
non_process
getitemquerystreamiterator api name casing and sample issues following the docs here i was expecting this code csharp public class todoactivity public string id get set public string status get set public int cost get set querydefinition querydefinition new querydefinition select from todos t where t cost expensive withparameter expensive await foreach response response in this container getitemquerystreamiterator querydefinition null new queryrequestoptions partitionkey new partitionkey error using streamreader sr new streamreader response content using jsontextreader jtr new jsontextreader sr jobject result jobject load jtr to return each document inside of the await foreach loop but it actually returns a structure like this documents so i had to create a class like this csharp public class querystream public queuemessage documents get set and then deserialize using that class issue documents casing because documents is pascal casing it will conflict with my serializationoptions and child documents which have camelcasing question should we use documents instead or enable the user to specify casing for the outer documents or ask users to name their property public queuemessage documents so the casing is honor as you can see above i needed to add a jsonpropertynames attribute because i m deserializing with camel and so it expects documents to be camel as well issue api name based on the example and the api name i was expecting it to yield for every document not every page question should this be renamed to getitempagequerystreamiterator issue docs are misleading because the doc has a single item class todoactivity i was led to believe it would return individual items suggestion update the sample to make it very clear what the dev has to do to use it here s the code to get this working this included a lot of trial and error and debugging to figure out csharp using system using system collections generic using system io using system text json using system text json serialization using system threading tasks using azure using azure cosmos using azure cosmos serialization using dotnetenv namespace cosmostest public class querystream public queuemessage documents get set class program static async task main string args env load cosmosclientoptions options new cosmosclientoptions serializeroptions new cosmosserializationoptions propertynamingpolicy cosmospropertynamingpolicy default cosmosclient cosmosclient new cosmosclient environment getenvironmentvariable azure cosmos endpoint environment getenvironmentvariable azure cosmos key options cosmoscontainer cosmoscontainer cosmosclient getdatabase environment getenvironmentvariable azure cosmos db getcontainer environment getenvironmentvariable azure cosmos container querydefinition querydefinition new querydefinition select from c list msgs new list await foreach response response in cosmoscontainer getitemquerystreamiterator querydefinition var querystream await jsonserializer deserializeasync response contentstream new jsonserializeroptions propertynamingpolicy jsonnamingpolicy camelcase msgs addrange querystream documents
0
12,802
15,181,113,785
IssuesEvent
2021-02-15 02:26:18
Geonovum/disgeo-arch
https://api.github.com/repos/Geonovum/disgeo-arch
closed
reactie gemeente Apeldoorn
Actoren Afbakening Cosmetisch Gegevenskwaliteit In Behandeling In behandeling - voorstel principes In behandeling - voorstel processen e.d. In behandeling - voorstel servicelayering Principes Processen Functies Componenten Service layering
1. Paragraaf 2.2, omschrijving processtap inwinnen, • Wat is het verschil tussen gegevens en eigenschappen? Het zinsdeel “over objecten en/of eigenschappen daarvan” is overbodig. Dat geldt voor alle keren dat dit genoemd wordt. Maakt het tegelijk leesbaarder. • Gegevensbron: Dat is dan toch een registratie? Wordt hier een andere registratie bedoeld dan de registratie bij het kopje Registreren? • De beschrijving gaat veel verder dan ‘inwinnen’. De tekst vanaf “beschikbaar maken…” hoort niet hier, omdat het niets te maken heeft met inwinnen 2. Paragraaf 2.2, omschrijving processtap samenstellen • Onduidelijk wie of wat hier iets doet • Onduidelijk wat een “samenhangende omschrijving” is en hoe dat eruit ziet 3. Paragraaf 2.2, omschrijving processtap registreren • Onduidelijk wat wordt bedoeld met en hoe dit eruit ziet: “op een gevalideerde wijze” • Waar ligt vast wie wat mag registreren? 4. Paragraaf 2.2, omschrijving processtap verrijken • Wie gaat dit doen? Voor wie ligt hier een rol? 5. Paragraaf 2.2, omschrijving processtap onderzoeken • Al dan niet wijzigen van het betreffende gegeven in de registratie’ kun je beter weer laten vallen onder inwinnen en de cirkel opnieuw starten. Tekst aanpassen: “… en na het verzamelen van aanvullende gegevens of indien er onjuistheden zijn opnieuw starten met ‘inwinnen’ of indien er geen onjuistheden zijn dit terugkoppelen aan de terugmelder.” 6. Paragraaf 2.3, tekst onder figuur 4 • ondersteundende = ondersteunende 7. Paragraaf 2.4 figuur 5 Interacties • Stippellijn van ‘Omgeving’ is zichtbaar, vermoedelijk staat onderin de stippellijn van ‘Bestuurd systeem’ deze stippellijn en de tekst ‘Bestuurd systeem’ is alleen niet zichtbaar in de afbeelding 8. Paragraaf 2.4, omschrijving bronhouder interactie gegevens • Wat betekent ‘beheert’? Is dat ‘onderhouden’? Of ‘veilig bewaren’? Of ….? Graag verduidelijken. • Is het gegeven pas authentiek als het in de SOR zit? Of als het in de eigen registratie zit (zoals nu bij de BAG). Graag een uitspraak hierover. 9. Paragraaf 2.4, omschrijving bronhouder interactie inzicht • Hoe bewaken we dan de kwaliteit? Hoe komt dat inzicht tot stand? 10. Paragraaf 2.4, bronhouder • Is de bronhouder automatisch afnemer? 11. Paragraaf 2.4, bronhouder interactie inzicht • Hoe komt het inzicht in gegevenskwaliteit tot stand? 12. Paragraaf 2.4, laatste alinea onder noot • Onduidelijk wie of wat de “Besturing van de Objectenregistratie” is. Wie is de eigenaar? 13. Paragraaf 3.2, laatste alinea • Wie zijn dienstenaanbieders? Software leveranciers? Of is dit de organisatie die de SOR (als centrale registratie) beheert? 14. Paragraaf 3.3, inrichtingsprincipe 1 • Door de tekst die achter de komma staat ontstaat er onduidelijkheid. De bronhouder heeft dus geen authentieke gegevens meer? 15. Paragraaf 3.3, inrichtingsprincipe 2 • Komt er één centrale registratie of wordt dit anders opgelost? Komt er toegang tot 1 registratie of 1 toegang tot meerdere registraties? 16. Paragraaf 3.3, inrichtingsprincipe 3 • Tekst “benaderen” in de eerste zin: Verwarrende term. Bedoel je raadplegen of registreren of beide? • Tekst “services” in de eerste zin: Raadpleegservices kunnen niet garanderen dat de gegevens voldoen aan de eisen. Dat kun je deels doen bij de registratie van gegevens door duidelijke bedrijfsregels toe te passen. 17. Paragraaf 3.3, inrichtingsprincipe 5 tekst “Alles is een service” • Vreemde uitspraak. Volgens mij niet correct. Voorbeeld: een gegeven is geen service. 18. Paragraaf 3.3, inrichtingsprincipe 5 tekst “Intern is extern” • Graag verduidelijken, dit is op meerdere manieren te lezen en nu niet duidelijk wat er bedoeld wordt 19. Paragraaf 4.2, tekst onder figuur 6 • “De laag Inzicht…” moet zijn “De kolom Inzicht…” 20. Paragraaf 4.3, omschrijving gegevenskwaliteit, tekst “de gemeten waarden” • Hoe doe je dat? Hoe ga je de kwaliteit van de gegevensverzameling meten en beschikbaar stellen? 21. Paragraaf 4.4 Functies in de laag Uitvoering • De functie Onderzoek ontbreekt. Komen daar ook faciliteiten voor? 22. Paragraaf 4.5 omschrijving betalingen • Wat zij betaalde diensten en wie bepaalt dit. Het kan niet zo zijn dat een bronhouder moeten betalen om de opgevoerde gegevens te gebruiken! 23. Paragraaf 4.7 Noot • Beschikbaarheid, performance, eenduidigheid van gebruik en inzetbaarheid van metadata 24. Paragraaf 5.2.1 1e alinea tekst “meta-gegevens” • Wat wordt verstaan onder meta-gegevens. Zijn dit de definities van de gegevenstypen of zijn dit bijvoorbeeld aanduidingen over de nauwkeurigheid van geregistreerde gegevens. 25. Paragraaf 5.2.1 tekst “De component Registratie biedt services” • Welke soorten services worden hier bedoeld en op basis waarvan? 26. Paragraaf 5.2.1 vereisen 1 • Is dit bij de SOR? Of bij de bronhouder die registreert? 27. Paragraaf 5.2.1 vereisen 2 • Wie legt gegevenswijzigingen vast? Welke gegevens worden bedoeld? Vastgelegd in de database of in een document? 28. Paragraaf 5.2.2 uitgangspunten, tekst “intern onderdeel” • Onduidelijk of intern bij de bronhouder is of in de SOR. Deze onduidelijkheid komt steeds terug. Maak een keuze! 29. Paragraaf 5.2.2 uitgangspunten, tekst “De technische wijze van opslag is verantwoordelijkheid van de uitvoeringspartij die dit invult.” • Is dit inwinnen, samenstellen en registreren of het beheren van de fysieke database. Toevoegen bij de stappen de uitvoeringspartijen. Wie is wie en wie doet wat? 30. Paragraaf 5.2.2 vereisten 2 • Om wat voor gegevens gaat het dan? De SOR bevat de in het woordenboek beschreven gegevens, niet andere gegevens die wellicht nodig zijn om de SOR-gegevens vast te stellen. Of gaat het hier bijvoorbeeld over gegevens voor autorisatie of logging? 31. Paragraaf 5.2.2 vereisten 6 • Wat bedoel wordt bedoeld met “opslagmechanisme”. Waarom spreken we nu over een opslagmechanisme. Beter punt naar overgezet kunnen worden. Nu absoluut niet duidelijk! Verduidelijken met voorbeelden? 32. Paragraaf 5.2.3, 2e alinea tekst “gegevens en informatie” • Niet mee eens. Kwestie van definiëren. Informatie is de presentatie van de gewenste gegevens in de juiste vorm op het juiste moment. Een afgeleid gegeven is dus nog geen informatie maar gewoon een gegeven. Het is dus nog maar de vraag of je met de SOR informatie wilt leveren. 33. Paragraaf 5.2.3, vereisten 2 • Onduidelijk is wat gemaks- en processervices zijn. 34. Paragraaf 5.2.3, vereisten 3 • Het zou makkelijk zijn als de SOR zelf pakt wat nodig is, dus schifting bij de SOR en niet bij de aanleverende partij. Maak gebruik van de dump en grap methode! 35. Paragraaf 5.2.3, noot • Welke andere soort services dan dataservices wordt dan aan gedacht? • Het onderscheid blijkt wel uit de definitie van de service. Een afnameservice biedt geen registratiemogelijkheden. Bij een registratieservice is het mogelijk dat je eerst gegevens moet opvragen, bijvoorbeeld om een relatie met een geregistreerd object te leggen. 36. Paragraaf 5.2.3.1 Afgeleide opslag • Door de afgeleide opslag worden bronhouders in een onmogelijke positie gebracht, omdat de bronhouder op het zelfde moment ook afnemer kan zijn 37. Paragraaf 5.2.3.1 vereisten • Kunnen we niet zonder die afgeleide opslag? Lijkt wat voorbarig om nu al over afgeleide opslag te praten. Liever de opslag zodanig structureren dat directe afname mogelijk is. Of speelt hier toch weer de onduidelijkheid over een centrale versus vele decentrale opslagplekken een rol? Is die afgeleide opslag dan een middel om de decentraal opgeslagen gegevens te verzamelen? 38. Paragraaf 5.2.3.1 vereisten 2 • Dus een kopie database bij de bronhouder? Hiermee haal je de complete inrichting van de SOR onderuit! 39. Paragraaf 5.2.3.1 vereisten 3 • Wie beheert/gebruikt de afgeleide opslag. Is dit onderdeel van de SOR of bij de afnemer of bronhouder? 40. Paragraaf 5.2.4 Notificatie • Mooi concept maar het vereist wel duidelijkheid over alle soorten gebeurtenissen waardoor gegevens kunnen wijzigen. Dat was bij het HR een flinke klus. 41. Paragraaf 5.2.4 Notificatie, invulling, tekst “Op een later moment wordt bepaald….” • Deze architectuurbeschrijving bevat nog veel open eindjes. 42. Paragraaf 5.2.5 uitgangspunten, 1e punt tekst “Bronhouders zijn geabonneerd…” • Abonneren suggereert vrijwilligheid. Dat is hier niet aan de orde. Bronhouders zijn verplicht terugmeldingen te onderzoeken. 43. Paragraaf 5.2.5 uitgangspunten, 2e punt tekst “terugmelding op een gegeven is zowel een aanduiding bij een gegeven dat er twijfel over bestaat als een aanleiding” • Even iets nauwkeuriger: een terugmelding is een gebeurtenis waarbij een twijfel over de juistheid van een geregistreerd gegeven wordt geuit. Deze terugmelding is tevens de aanleiding voor de bronhouder om die twijfel te onderzoeken. 44. Paragraaf 5.2.5 uitgangspunten, 4e punt tekst “standaard” • Welke standaard gegevens? 45. Paragraaf 5.2.5 uitgangspunten, 6e punt tekst “betrekking hebben op 1 of meerdere bronhouders.” • Zijn er dan meerdere bronhouders voor 1 gegeven? 46. Paragraaf 5.2.5 vereisten 2 tekst “bij registratie” • Dan is het toch geen terugmelding? Je gaat een gegeven toch niet registreren als het niet strookt met de beperkingsregels? 47. Paragraaf 5.2.5 vereisten 2 tekst “periodiek” • Niet achteraf maar vooraf dwz voor registratie controleren! 48. Paragraaf 5.2.5 vereisten 2 tekst “veranderingen in de gegevensstructuur” • Hoe kan een verandering in de gegevensstructuur leiden tot twijfels over de juistheid? 49. Paragraaf 5.2.5 vereisten 4 tekst “bulk” • Dat leidt dan wel tot evenveel terugmeldingen als er gegevens zijn waarover getwijfeld wordt? Immers iedere terugmelding wordt gerelateerd aan het betreffende geregistreerde gegeven. 50. Paragraaf 5.2.5 Externe afhankelijkheden • Wat doe je met die afhankelijkheid? 51. Paragraaf 5.3.1 1e zin tekst “informatieproducten” • Zoals eerder aangegeven: een afgeleid gegeven is een gegeven en geen informatie. Goed om afgeleide gegevens op te nemen in de Gegevenscatalogus, graag met afleidingsregel. 52. Paragraaf 5.3.1 2e zin tekst “inhoud van de gegevens en informatieproducten” • Niet de inhoud maar de betekenis van gegevens. • Beter zicht nodig op waaraan bij informatieproducten gedacht wordt. 53. Paragraaf 5.3.1 4e zin tekst “aanname” • Schrijft de architectuur dat niet voor? Dit is toch een verplichting? 54. Paragraaf 5.3.1 vereisten • De beperkingsregels zoals optionaliteit, kardinaliteit ontbreken 55. Paragraaf 5.3.1 vereisten 1 tekst “begrippen” • Vervangen door objecttypen en gegevenstypen. 56. Paragraaf 5.3.1 vereisten 2 tekst “relaties tussen de begrippen” • Graag uitleg wat hier bedoeld wordt. Gaat het over begrippen of over gegevens. En wat voor relaties? Worden hier beperkingsregels bedoeld? Bijvoorbeeld een Pand bevat 0, 1 of meer verblijfsobjecten? 57. Paragraaf 5.3.1 vereisten 5 tekst “veranderingen van relaties, verandering van definities” • Mogelijke veranderingen hebben niet onze voorkeur. 58. Paragraaf 5.3.1 vereisten 8 tekst “inhoud” • Of betekenis? 59. Paragraaf 5.3.2 vereisten 1 • Wie gaat dit doen? Op welke plek worden ze ingezet? 60. Paragraaf 5.3.2 vereisten 6.3 • ???? 61. Paragraaf 5.3.3 tekst “(structuur van de) inhoud” • “(structuur van de inhoud)” voegt niet toe, kan weg • Inhoud wijzigen in betekenis 62. Paragraaf 5.4.1, 2e zin tekst “abonnementen” • Dus niet van toepassing op terugmeldingen 63. Paragraaf 5.4.1.1 tekst “Abonnementen zijn abonnementen” • Tekst wijzigen in “Notificatieabonnementen zijn abonnementen” 64. Paragraaf 5.4.1.2 noot • Services moeten ook zonder abonnement beschikbaar zijn! En door de bronhouder te allen tijde kosteloos gebruikt kunnen worden! 65. Paragraaf 5.4.1.2 uitgangspunten, 2x 1e woord Abonnementen • Wijzigen in afnameabonnementen 66. Paragraaf 5.5.1 uitgangspunten • Worden informatiesystemen ook gecertificeerd? 67. Paragraaf 5.5.2 Invulling • Beter om de twee punten te plaatsen onder Vereisten 68. Paragraaf 6.1 regel 05 • Veel te algemeen. Beschrijf wat je bedoelt. 69. Paragraaf 6.6 regel 10, tekst “Informatie is “bewerkte data”” • Niet geheel mee eens. Een afgeleid gegeven is ook een gegeven. Daar kan kennis, bijvoorbeeld in de vorm van een rekenregel, aan zijn toegevoegd. Het is pas informatie als het op het juiste moment in de juiste vorm wordt gepresenteerd.
2.0
reactie gemeente Apeldoorn - 1. Paragraaf 2.2, omschrijving processtap inwinnen, • Wat is het verschil tussen gegevens en eigenschappen? Het zinsdeel “over objecten en/of eigenschappen daarvan” is overbodig. Dat geldt voor alle keren dat dit genoemd wordt. Maakt het tegelijk leesbaarder. • Gegevensbron: Dat is dan toch een registratie? Wordt hier een andere registratie bedoeld dan de registratie bij het kopje Registreren? • De beschrijving gaat veel verder dan ‘inwinnen’. De tekst vanaf “beschikbaar maken…” hoort niet hier, omdat het niets te maken heeft met inwinnen 2. Paragraaf 2.2, omschrijving processtap samenstellen • Onduidelijk wie of wat hier iets doet • Onduidelijk wat een “samenhangende omschrijving” is en hoe dat eruit ziet 3. Paragraaf 2.2, omschrijving processtap registreren • Onduidelijk wat wordt bedoeld met en hoe dit eruit ziet: “op een gevalideerde wijze” • Waar ligt vast wie wat mag registreren? 4. Paragraaf 2.2, omschrijving processtap verrijken • Wie gaat dit doen? Voor wie ligt hier een rol? 5. Paragraaf 2.2, omschrijving processtap onderzoeken • Al dan niet wijzigen van het betreffende gegeven in de registratie’ kun je beter weer laten vallen onder inwinnen en de cirkel opnieuw starten. Tekst aanpassen: “… en na het verzamelen van aanvullende gegevens of indien er onjuistheden zijn opnieuw starten met ‘inwinnen’ of indien er geen onjuistheden zijn dit terugkoppelen aan de terugmelder.” 6. Paragraaf 2.3, tekst onder figuur 4 • ondersteundende = ondersteunende 7. Paragraaf 2.4 figuur 5 Interacties • Stippellijn van ‘Omgeving’ is zichtbaar, vermoedelijk staat onderin de stippellijn van ‘Bestuurd systeem’ deze stippellijn en de tekst ‘Bestuurd systeem’ is alleen niet zichtbaar in de afbeelding 8. Paragraaf 2.4, omschrijving bronhouder interactie gegevens • Wat betekent ‘beheert’? Is dat ‘onderhouden’? Of ‘veilig bewaren’? Of ….? Graag verduidelijken. • Is het gegeven pas authentiek als het in de SOR zit? Of als het in de eigen registratie zit (zoals nu bij de BAG). Graag een uitspraak hierover. 9. Paragraaf 2.4, omschrijving bronhouder interactie inzicht • Hoe bewaken we dan de kwaliteit? Hoe komt dat inzicht tot stand? 10. Paragraaf 2.4, bronhouder • Is de bronhouder automatisch afnemer? 11. Paragraaf 2.4, bronhouder interactie inzicht • Hoe komt het inzicht in gegevenskwaliteit tot stand? 12. Paragraaf 2.4, laatste alinea onder noot • Onduidelijk wie of wat de “Besturing van de Objectenregistratie” is. Wie is de eigenaar? 13. Paragraaf 3.2, laatste alinea • Wie zijn dienstenaanbieders? Software leveranciers? Of is dit de organisatie die de SOR (als centrale registratie) beheert? 14. Paragraaf 3.3, inrichtingsprincipe 1 • Door de tekst die achter de komma staat ontstaat er onduidelijkheid. De bronhouder heeft dus geen authentieke gegevens meer? 15. Paragraaf 3.3, inrichtingsprincipe 2 • Komt er één centrale registratie of wordt dit anders opgelost? Komt er toegang tot 1 registratie of 1 toegang tot meerdere registraties? 16. Paragraaf 3.3, inrichtingsprincipe 3 • Tekst “benaderen” in de eerste zin: Verwarrende term. Bedoel je raadplegen of registreren of beide? • Tekst “services” in de eerste zin: Raadpleegservices kunnen niet garanderen dat de gegevens voldoen aan de eisen. Dat kun je deels doen bij de registratie van gegevens door duidelijke bedrijfsregels toe te passen. 17. Paragraaf 3.3, inrichtingsprincipe 5 tekst “Alles is een service” • Vreemde uitspraak. Volgens mij niet correct. Voorbeeld: een gegeven is geen service. 18. Paragraaf 3.3, inrichtingsprincipe 5 tekst “Intern is extern” • Graag verduidelijken, dit is op meerdere manieren te lezen en nu niet duidelijk wat er bedoeld wordt 19. Paragraaf 4.2, tekst onder figuur 6 • “De laag Inzicht…” moet zijn “De kolom Inzicht…” 20. Paragraaf 4.3, omschrijving gegevenskwaliteit, tekst “de gemeten waarden” • Hoe doe je dat? Hoe ga je de kwaliteit van de gegevensverzameling meten en beschikbaar stellen? 21. Paragraaf 4.4 Functies in de laag Uitvoering • De functie Onderzoek ontbreekt. Komen daar ook faciliteiten voor? 22. Paragraaf 4.5 omschrijving betalingen • Wat zij betaalde diensten en wie bepaalt dit. Het kan niet zo zijn dat een bronhouder moeten betalen om de opgevoerde gegevens te gebruiken! 23. Paragraaf 4.7 Noot • Beschikbaarheid, performance, eenduidigheid van gebruik en inzetbaarheid van metadata 24. Paragraaf 5.2.1 1e alinea tekst “meta-gegevens” • Wat wordt verstaan onder meta-gegevens. Zijn dit de definities van de gegevenstypen of zijn dit bijvoorbeeld aanduidingen over de nauwkeurigheid van geregistreerde gegevens. 25. Paragraaf 5.2.1 tekst “De component Registratie biedt services” • Welke soorten services worden hier bedoeld en op basis waarvan? 26. Paragraaf 5.2.1 vereisen 1 • Is dit bij de SOR? Of bij de bronhouder die registreert? 27. Paragraaf 5.2.1 vereisen 2 • Wie legt gegevenswijzigingen vast? Welke gegevens worden bedoeld? Vastgelegd in de database of in een document? 28. Paragraaf 5.2.2 uitgangspunten, tekst “intern onderdeel” • Onduidelijk of intern bij de bronhouder is of in de SOR. Deze onduidelijkheid komt steeds terug. Maak een keuze! 29. Paragraaf 5.2.2 uitgangspunten, tekst “De technische wijze van opslag is verantwoordelijkheid van de uitvoeringspartij die dit invult.” • Is dit inwinnen, samenstellen en registreren of het beheren van de fysieke database. Toevoegen bij de stappen de uitvoeringspartijen. Wie is wie en wie doet wat? 30. Paragraaf 5.2.2 vereisten 2 • Om wat voor gegevens gaat het dan? De SOR bevat de in het woordenboek beschreven gegevens, niet andere gegevens die wellicht nodig zijn om de SOR-gegevens vast te stellen. Of gaat het hier bijvoorbeeld over gegevens voor autorisatie of logging? 31. Paragraaf 5.2.2 vereisten 6 • Wat bedoel wordt bedoeld met “opslagmechanisme”. Waarom spreken we nu over een opslagmechanisme. Beter punt naar overgezet kunnen worden. Nu absoluut niet duidelijk! Verduidelijken met voorbeelden? 32. Paragraaf 5.2.3, 2e alinea tekst “gegevens en informatie” • Niet mee eens. Kwestie van definiëren. Informatie is de presentatie van de gewenste gegevens in de juiste vorm op het juiste moment. Een afgeleid gegeven is dus nog geen informatie maar gewoon een gegeven. Het is dus nog maar de vraag of je met de SOR informatie wilt leveren. 33. Paragraaf 5.2.3, vereisten 2 • Onduidelijk is wat gemaks- en processervices zijn. 34. Paragraaf 5.2.3, vereisten 3 • Het zou makkelijk zijn als de SOR zelf pakt wat nodig is, dus schifting bij de SOR en niet bij de aanleverende partij. Maak gebruik van de dump en grap methode! 35. Paragraaf 5.2.3, noot • Welke andere soort services dan dataservices wordt dan aan gedacht? • Het onderscheid blijkt wel uit de definitie van de service. Een afnameservice biedt geen registratiemogelijkheden. Bij een registratieservice is het mogelijk dat je eerst gegevens moet opvragen, bijvoorbeeld om een relatie met een geregistreerd object te leggen. 36. Paragraaf 5.2.3.1 Afgeleide opslag • Door de afgeleide opslag worden bronhouders in een onmogelijke positie gebracht, omdat de bronhouder op het zelfde moment ook afnemer kan zijn 37. Paragraaf 5.2.3.1 vereisten • Kunnen we niet zonder die afgeleide opslag? Lijkt wat voorbarig om nu al over afgeleide opslag te praten. Liever de opslag zodanig structureren dat directe afname mogelijk is. Of speelt hier toch weer de onduidelijkheid over een centrale versus vele decentrale opslagplekken een rol? Is die afgeleide opslag dan een middel om de decentraal opgeslagen gegevens te verzamelen? 38. Paragraaf 5.2.3.1 vereisten 2 • Dus een kopie database bij de bronhouder? Hiermee haal je de complete inrichting van de SOR onderuit! 39. Paragraaf 5.2.3.1 vereisten 3 • Wie beheert/gebruikt de afgeleide opslag. Is dit onderdeel van de SOR of bij de afnemer of bronhouder? 40. Paragraaf 5.2.4 Notificatie • Mooi concept maar het vereist wel duidelijkheid over alle soorten gebeurtenissen waardoor gegevens kunnen wijzigen. Dat was bij het HR een flinke klus. 41. Paragraaf 5.2.4 Notificatie, invulling, tekst “Op een later moment wordt bepaald….” • Deze architectuurbeschrijving bevat nog veel open eindjes. 42. Paragraaf 5.2.5 uitgangspunten, 1e punt tekst “Bronhouders zijn geabonneerd…” • Abonneren suggereert vrijwilligheid. Dat is hier niet aan de orde. Bronhouders zijn verplicht terugmeldingen te onderzoeken. 43. Paragraaf 5.2.5 uitgangspunten, 2e punt tekst “terugmelding op een gegeven is zowel een aanduiding bij een gegeven dat er twijfel over bestaat als een aanleiding” • Even iets nauwkeuriger: een terugmelding is een gebeurtenis waarbij een twijfel over de juistheid van een geregistreerd gegeven wordt geuit. Deze terugmelding is tevens de aanleiding voor de bronhouder om die twijfel te onderzoeken. 44. Paragraaf 5.2.5 uitgangspunten, 4e punt tekst “standaard” • Welke standaard gegevens? 45. Paragraaf 5.2.5 uitgangspunten, 6e punt tekst “betrekking hebben op 1 of meerdere bronhouders.” • Zijn er dan meerdere bronhouders voor 1 gegeven? 46. Paragraaf 5.2.5 vereisten 2 tekst “bij registratie” • Dan is het toch geen terugmelding? Je gaat een gegeven toch niet registreren als het niet strookt met de beperkingsregels? 47. Paragraaf 5.2.5 vereisten 2 tekst “periodiek” • Niet achteraf maar vooraf dwz voor registratie controleren! 48. Paragraaf 5.2.5 vereisten 2 tekst “veranderingen in de gegevensstructuur” • Hoe kan een verandering in de gegevensstructuur leiden tot twijfels over de juistheid? 49. Paragraaf 5.2.5 vereisten 4 tekst “bulk” • Dat leidt dan wel tot evenveel terugmeldingen als er gegevens zijn waarover getwijfeld wordt? Immers iedere terugmelding wordt gerelateerd aan het betreffende geregistreerde gegeven. 50. Paragraaf 5.2.5 Externe afhankelijkheden • Wat doe je met die afhankelijkheid? 51. Paragraaf 5.3.1 1e zin tekst “informatieproducten” • Zoals eerder aangegeven: een afgeleid gegeven is een gegeven en geen informatie. Goed om afgeleide gegevens op te nemen in de Gegevenscatalogus, graag met afleidingsregel. 52. Paragraaf 5.3.1 2e zin tekst “inhoud van de gegevens en informatieproducten” • Niet de inhoud maar de betekenis van gegevens. • Beter zicht nodig op waaraan bij informatieproducten gedacht wordt. 53. Paragraaf 5.3.1 4e zin tekst “aanname” • Schrijft de architectuur dat niet voor? Dit is toch een verplichting? 54. Paragraaf 5.3.1 vereisten • De beperkingsregels zoals optionaliteit, kardinaliteit ontbreken 55. Paragraaf 5.3.1 vereisten 1 tekst “begrippen” • Vervangen door objecttypen en gegevenstypen. 56. Paragraaf 5.3.1 vereisten 2 tekst “relaties tussen de begrippen” • Graag uitleg wat hier bedoeld wordt. Gaat het over begrippen of over gegevens. En wat voor relaties? Worden hier beperkingsregels bedoeld? Bijvoorbeeld een Pand bevat 0, 1 of meer verblijfsobjecten? 57. Paragraaf 5.3.1 vereisten 5 tekst “veranderingen van relaties, verandering van definities” • Mogelijke veranderingen hebben niet onze voorkeur. 58. Paragraaf 5.3.1 vereisten 8 tekst “inhoud” • Of betekenis? 59. Paragraaf 5.3.2 vereisten 1 • Wie gaat dit doen? Op welke plek worden ze ingezet? 60. Paragraaf 5.3.2 vereisten 6.3 • ???? 61. Paragraaf 5.3.3 tekst “(structuur van de) inhoud” • “(structuur van de inhoud)” voegt niet toe, kan weg • Inhoud wijzigen in betekenis 62. Paragraaf 5.4.1, 2e zin tekst “abonnementen” • Dus niet van toepassing op terugmeldingen 63. Paragraaf 5.4.1.1 tekst “Abonnementen zijn abonnementen” • Tekst wijzigen in “Notificatieabonnementen zijn abonnementen” 64. Paragraaf 5.4.1.2 noot • Services moeten ook zonder abonnement beschikbaar zijn! En door de bronhouder te allen tijde kosteloos gebruikt kunnen worden! 65. Paragraaf 5.4.1.2 uitgangspunten, 2x 1e woord Abonnementen • Wijzigen in afnameabonnementen 66. Paragraaf 5.5.1 uitgangspunten • Worden informatiesystemen ook gecertificeerd? 67. Paragraaf 5.5.2 Invulling • Beter om de twee punten te plaatsen onder Vereisten 68. Paragraaf 6.1 regel 05 • Veel te algemeen. Beschrijf wat je bedoelt. 69. Paragraaf 6.6 regel 10, tekst “Informatie is “bewerkte data”” • Niet geheel mee eens. Een afgeleid gegeven is ook een gegeven. Daar kan kennis, bijvoorbeeld in de vorm van een rekenregel, aan zijn toegevoegd. Het is pas informatie als het op het juiste moment in de juiste vorm wordt gepresenteerd.
process
reactie gemeente apeldoorn paragraaf omschrijving processtap inwinnen • wat is het verschil tussen gegevens en eigenschappen het zinsdeel “over objecten en of eigenschappen daarvan” is overbodig dat geldt voor alle keren dat dit genoemd wordt maakt het tegelijk leesbaarder • gegevensbron dat is dan toch een registratie wordt hier een andere registratie bedoeld dan de registratie bij het kopje registreren • de beschrijving gaat veel verder dan ‘inwinnen’ de tekst vanaf “beschikbaar maken…” hoort niet hier omdat het niets te maken heeft met inwinnen paragraaf omschrijving processtap samenstellen • onduidelijk wie of wat hier iets doet • onduidelijk wat een “samenhangende omschrijving” is en hoe dat eruit ziet paragraaf omschrijving processtap registreren • onduidelijk wat wordt bedoeld met en hoe dit eruit ziet “op een gevalideerde wijze” • waar ligt vast wie wat mag registreren paragraaf omschrijving processtap verrijken • wie gaat dit doen voor wie ligt hier een rol paragraaf omschrijving processtap onderzoeken • al dan niet wijzigen van het betreffende gegeven in de registratie’ kun je beter weer laten vallen onder inwinnen en de cirkel opnieuw starten tekst aanpassen “… en na het verzamelen van aanvullende gegevens of indien er onjuistheden zijn opnieuw starten met ‘inwinnen’ of indien er geen onjuistheden zijn dit terugkoppelen aan de terugmelder ” paragraaf tekst onder figuur • ondersteundende ondersteunende paragraaf figuur interacties • stippellijn van ‘omgeving’ is zichtbaar vermoedelijk staat onderin de stippellijn van ‘bestuurd systeem’ deze stippellijn en de tekst ‘bestuurd systeem’ is alleen niet zichtbaar in de afbeelding paragraaf omschrijving bronhouder interactie gegevens • wat betekent ‘beheert’ is dat ‘onderhouden’ of ‘veilig bewaren’ of … graag verduidelijken • is het gegeven pas authentiek als het in de sor zit of als het in de eigen registratie zit zoals nu bij de bag graag een uitspraak hierover paragraaf omschrijving bronhouder interactie inzicht • hoe bewaken we dan de kwaliteit hoe komt dat inzicht tot stand paragraaf bronhouder • is de bronhouder automatisch afnemer paragraaf bronhouder interactie inzicht • hoe komt het inzicht in gegevenskwaliteit tot stand paragraaf laatste alinea onder noot • onduidelijk wie of wat de “besturing van de objectenregistratie” is wie is de eigenaar paragraaf laatste alinea • wie zijn dienstenaanbieders software leveranciers of is dit de organisatie die de sor als centrale registratie beheert paragraaf inrichtingsprincipe • door de tekst die achter de komma staat ontstaat er onduidelijkheid de bronhouder heeft dus geen authentieke gegevens meer paragraaf inrichtingsprincipe • komt er één centrale registratie of wordt dit anders opgelost komt er toegang tot registratie of toegang tot meerdere registraties paragraaf inrichtingsprincipe • tekst “benaderen” in de eerste zin verwarrende term bedoel je raadplegen of registreren of beide • tekst “services” in de eerste zin raadpleegservices kunnen niet garanderen dat de gegevens voldoen aan de eisen dat kun je deels doen bij de registratie van gegevens door duidelijke bedrijfsregels toe te passen paragraaf inrichtingsprincipe tekst “alles is een service” • vreemde uitspraak volgens mij niet correct voorbeeld een gegeven is geen service paragraaf inrichtingsprincipe tekst “intern is extern” • graag verduidelijken dit is op meerdere manieren te lezen en nu niet duidelijk wat er bedoeld wordt paragraaf tekst onder figuur • “de laag inzicht…” moet zijn “de kolom inzicht…” paragraaf omschrijving gegevenskwaliteit tekst “de gemeten waarden” • hoe doe je dat hoe ga je de kwaliteit van de gegevensverzameling meten en beschikbaar stellen paragraaf functies in de laag uitvoering • de functie onderzoek ontbreekt komen daar ook faciliteiten voor paragraaf omschrijving betalingen • wat zij betaalde diensten en wie bepaalt dit het kan niet zo zijn dat een bronhouder moeten betalen om de opgevoerde gegevens te gebruiken paragraaf noot • beschikbaarheid performance eenduidigheid van gebruik en inzetbaarheid van metadata paragraaf alinea tekst “meta gegevens” • wat wordt verstaan onder meta gegevens zijn dit de definities van de gegevenstypen of zijn dit bijvoorbeeld aanduidingen over de nauwkeurigheid van geregistreerde gegevens paragraaf tekst “de component registratie biedt services” • welke soorten services worden hier bedoeld en op basis waarvan paragraaf vereisen • is dit bij de sor of bij de bronhouder die registreert paragraaf vereisen • wie legt gegevenswijzigingen vast welke gegevens worden bedoeld vastgelegd in de database of in een document paragraaf uitgangspunten tekst “intern onderdeel” • onduidelijk of intern bij de bronhouder is of in de sor deze onduidelijkheid komt steeds terug maak een keuze paragraaf uitgangspunten tekst “de technische wijze van opslag is verantwoordelijkheid van de uitvoeringspartij die dit invult ” • is dit inwinnen samenstellen en registreren of het beheren van de fysieke database toevoegen bij de stappen de uitvoeringspartijen wie is wie en wie doet wat paragraaf vereisten • om wat voor gegevens gaat het dan de sor bevat de in het woordenboek beschreven gegevens niet andere gegevens die wellicht nodig zijn om de sor gegevens vast te stellen of gaat het hier bijvoorbeeld over gegevens voor autorisatie of logging paragraaf vereisten • wat bedoel wordt bedoeld met “opslagmechanisme” waarom spreken we nu over een opslagmechanisme beter punt naar overgezet kunnen worden nu absoluut niet duidelijk verduidelijken met voorbeelden paragraaf alinea tekst “gegevens en informatie” • niet mee eens kwestie van definiëren informatie is de presentatie van de gewenste gegevens in de juiste vorm op het juiste moment een afgeleid gegeven is dus nog geen informatie maar gewoon een gegeven het is dus nog maar de vraag of je met de sor informatie wilt leveren paragraaf vereisten • onduidelijk is wat gemaks en processervices zijn paragraaf vereisten • het zou makkelijk zijn als de sor zelf pakt wat nodig is dus schifting bij de sor en niet bij de aanleverende partij maak gebruik van de dump en grap methode paragraaf noot • welke andere soort services dan dataservices wordt dan aan gedacht • het onderscheid blijkt wel uit de definitie van de service een afnameservice biedt geen registratiemogelijkheden bij een registratieservice is het mogelijk dat je eerst gegevens moet opvragen bijvoorbeeld om een relatie met een geregistreerd object te leggen paragraaf afgeleide opslag • door de afgeleide opslag worden bronhouders in een onmogelijke positie gebracht omdat de bronhouder op het zelfde moment ook afnemer kan zijn paragraaf vereisten • kunnen we niet zonder die afgeleide opslag lijkt wat voorbarig om nu al over afgeleide opslag te praten liever de opslag zodanig structureren dat directe afname mogelijk is of speelt hier toch weer de onduidelijkheid over een centrale versus vele decentrale opslagplekken een rol is die afgeleide opslag dan een middel om de decentraal opgeslagen gegevens te verzamelen paragraaf vereisten • dus een kopie database bij de bronhouder hiermee haal je de complete inrichting van de sor onderuit paragraaf vereisten • wie beheert gebruikt de afgeleide opslag is dit onderdeel van de sor of bij de afnemer of bronhouder paragraaf notificatie • mooi concept maar het vereist wel duidelijkheid over alle soorten gebeurtenissen waardoor gegevens kunnen wijzigen dat was bij het hr een flinke klus paragraaf notificatie invulling tekst “op een later moment wordt bepaald… ” • deze architectuurbeschrijving bevat nog veel open eindjes paragraaf uitgangspunten punt tekst “bronhouders zijn geabonneerd…” • abonneren suggereert vrijwilligheid dat is hier niet aan de orde bronhouders zijn verplicht terugmeldingen te onderzoeken paragraaf uitgangspunten punt tekst “terugmelding op een gegeven is zowel een aanduiding bij een gegeven dat er twijfel over bestaat als een aanleiding” • even iets nauwkeuriger een terugmelding is een gebeurtenis waarbij een twijfel over de juistheid van een geregistreerd gegeven wordt geuit deze terugmelding is tevens de aanleiding voor de bronhouder om die twijfel te onderzoeken paragraaf uitgangspunten punt tekst “standaard” • welke standaard gegevens paragraaf uitgangspunten punt tekst “betrekking hebben op of meerdere bronhouders ” • zijn er dan meerdere bronhouders voor gegeven paragraaf vereisten tekst “bij registratie” • dan is het toch geen terugmelding je gaat een gegeven toch niet registreren als het niet strookt met de beperkingsregels paragraaf vereisten tekst “periodiek” • niet achteraf maar vooraf dwz voor registratie controleren paragraaf vereisten tekst “veranderingen in de gegevensstructuur” • hoe kan een verandering in de gegevensstructuur leiden tot twijfels over de juistheid paragraaf vereisten tekst “bulk” • dat leidt dan wel tot evenveel terugmeldingen als er gegevens zijn waarover getwijfeld wordt immers iedere terugmelding wordt gerelateerd aan het betreffende geregistreerde gegeven paragraaf externe afhankelijkheden • wat doe je met die afhankelijkheid paragraaf zin tekst “informatieproducten” • zoals eerder aangegeven een afgeleid gegeven is een gegeven en geen informatie goed om afgeleide gegevens op te nemen in de gegevenscatalogus graag met afleidingsregel paragraaf zin tekst “inhoud van de gegevens en informatieproducten” • niet de inhoud maar de betekenis van gegevens • beter zicht nodig op waaraan bij informatieproducten gedacht wordt paragraaf zin tekst “aanname” • schrijft de architectuur dat niet voor dit is toch een verplichting paragraaf vereisten • de beperkingsregels zoals optionaliteit kardinaliteit ontbreken paragraaf vereisten tekst “begrippen” • vervangen door objecttypen en gegevenstypen paragraaf vereisten tekst “relaties tussen de begrippen” • graag uitleg wat hier bedoeld wordt gaat het over begrippen of over gegevens en wat voor relaties worden hier beperkingsregels bedoeld bijvoorbeeld een pand bevat of meer verblijfsobjecten paragraaf vereisten tekst “veranderingen van relaties verandering van definities” • mogelijke veranderingen hebben niet onze voorkeur paragraaf vereisten tekst “inhoud” • of betekenis paragraaf vereisten • wie gaat dit doen op welke plek worden ze ingezet paragraaf vereisten • paragraaf tekst “ structuur van de inhoud” • “ structuur van de inhoud ” voegt niet toe kan weg • inhoud wijzigen in betekenis paragraaf zin tekst “abonnementen” • dus niet van toepassing op terugmeldingen paragraaf tekst “abonnementen zijn abonnementen” • tekst wijzigen in “notificatieabonnementen zijn abonnementen” paragraaf noot • services moeten ook zonder abonnement beschikbaar zijn en door de bronhouder te allen tijde kosteloos gebruikt kunnen worden paragraaf uitgangspunten woord abonnementen • wijzigen in afnameabonnementen paragraaf uitgangspunten • worden informatiesystemen ook gecertificeerd paragraaf invulling • beter om de twee punten te plaatsen onder vereisten paragraaf regel • veel te algemeen beschrijf wat je bedoelt paragraaf regel tekst “informatie is “bewerkte data”” • niet geheel mee eens een afgeleid gegeven is ook een gegeven daar kan kennis bijvoorbeeld in de vorm van een rekenregel aan zijn toegevoegd het is pas informatie als het op het juiste moment in de juiste vorm wordt gepresenteerd
1
198,938
15,728,806,653
IssuesEvent
2021-03-29 14:12:23
nautobot/nautobot
https://api.github.com/repos/nautobot/nautobot
closed
Remove CSRF_TRUSTED_ORIGINS from core settings
status: current type: bug type: documentation
<!-- NOTE: This template is for use by maintainers only. Please do not submit an issue using this template unless you have been specifically asked to do so. --> ### Proposed Changes Remove `CSRF_TRUSTED_ORIGINS` from `nautobot.core.settings`. Replace it with revised mention of this setting in the documentation. Currently it is mentioned as being set to the same value as `ALLOWED_HOSTS` which is no longer possible due to the fact that all those values are defaults and any settings overloaded in a user's `nautobot_config.py` will not be evaluated again by values already set in `nautobot.core.settings`. <!-- Provide justification for the proposed change(s). --> ### Justification This is a "dynamic" setting from NetBox that has been eliminated in Nautobot. It is no longer accurate as documented.
1.0
Remove CSRF_TRUSTED_ORIGINS from core settings - <!-- NOTE: This template is for use by maintainers only. Please do not submit an issue using this template unless you have been specifically asked to do so. --> ### Proposed Changes Remove `CSRF_TRUSTED_ORIGINS` from `nautobot.core.settings`. Replace it with revised mention of this setting in the documentation. Currently it is mentioned as being set to the same value as `ALLOWED_HOSTS` which is no longer possible due to the fact that all those values are defaults and any settings overloaded in a user's `nautobot_config.py` will not be evaluated again by values already set in `nautobot.core.settings`. <!-- Provide justification for the proposed change(s). --> ### Justification This is a "dynamic" setting from NetBox that has been eliminated in Nautobot. It is no longer accurate as documented.
non_process
remove csrf trusted origins from core settings note this template is for use by maintainers only please do not submit an issue using this template unless you have been specifically asked to do so proposed changes remove csrf trusted origins from nautobot core settings replace it with revised mention of this setting in the documentation currently it is mentioned as being set to the same value as allowed hosts which is no longer possible due to the fact that all those values are defaults and any settings overloaded in a user s nautobot config py will not be evaluated again by values already set in nautobot core settings justification this is a dynamic setting from netbox that has been eliminated in nautobot it is no longer accurate as documented
0
223,550
17,606,194,438
IssuesEvent
2021-08-17 17:23:49
nasa/cFE
https://api.github.com/repos/nasa/cFE
closed
Write Resource ID functional test to check maximum
unit-test
**Is your feature request related to a problem? Please describe.** Need open source functional tests for certifiability **Describe the solution you'd like** Add additional checks to functional tests for ES Resource ID to Index that ensure the returned index is less than the maximum allowed for that resource type **Describe alternatives you've considered** None **Requester Info** _Originally posted by @astrogeco in https://github.com/nasa/cFE/pull/1734#issuecomment-892799313_
1.0
Write Resource ID functional test to check maximum - **Is your feature request related to a problem? Please describe.** Need open source functional tests for certifiability **Describe the solution you'd like** Add additional checks to functional tests for ES Resource ID to Index that ensure the returned index is less than the maximum allowed for that resource type **Describe alternatives you've considered** None **Requester Info** _Originally posted by @astrogeco in https://github.com/nasa/cFE/pull/1734#issuecomment-892799313_
non_process
write resource id functional test to check maximum is your feature request related to a problem please describe need open source functional tests for certifiability describe the solution you d like add additional checks to functional tests for es resource id to index that ensure the returned index is less than the maximum allowed for that resource type describe alternatives you ve considered none requester info originally posted by astrogeco in
0
19,408
25,553,881,401
IssuesEvent
2022-11-30 03:48:17
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Add Spotless to Gradle
enhancement process
### Problem We need to reach feature parity with the Maven build before we can replace it with Gradle. ### Solution * Add Spotless plugin to Gradle * Add license header updating to spotless * Add version replacement functionality ### Alternatives _No response_
1.0
Add Spotless to Gradle - ### Problem We need to reach feature parity with the Maven build before we can replace it with Gradle. ### Solution * Add Spotless plugin to Gradle * Add license header updating to spotless * Add version replacement functionality ### Alternatives _No response_
process
add spotless to gradle problem we need to reach feature parity with the maven build before we can replace it with gradle solution add spotless plugin to gradle add license header updating to spotless add version replacement functionality alternatives no response
1
19,275
25,463,982,928
IssuesEvent
2022-11-25 00:39:21
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
Analista de Qualidade na [XTESTING]
SALVADOR EFETIVO(CLT) SCRUM MPS.BR ISO 9001 CMMI MPT.BR QUALIDADE MODELAGEM DE PROCESSOS Stale
### Analista de Qualidade Vaga para analista de qualidade na XTESTING ## Atividades Profissional responsável por atuar com autorias internas e externas, fazendo a elaboração de normas e procedimento e não conformidades dos processos de qualidade que se relaciona com todas as áreas ligadas a qualidade dentro de uma empresa. ## Local - Salvador ## Benefícios - Transporte; - Ticket Refeição; - Assistência médica; ## Requisitos **Obrigatórios:** - Nível superior completo; - Ter participado de projeto na área de qualidade; - Modelagem de processos; - Metodologias; - ISO 9001; **Diferenciais:** - MPT.Br; - CMMI; - MPS.Br; - SCRUM; ## Contratação - CLT; - Salário a combinar; ## XTESTING Fundada em 2013, a X-Testing é uma empresa especializada na prestação de serviços na área de teste de software com experiência na implantação de testes funcionais, de performance, usabilidade, segurança e acessibilidade em empresas privadas e do setor público. Promove a qualidade dos softwares desenvolvidos através de processos e metodologias certificados pelo modelo MPT. BR nível 4, utilizando as melhores ferramentas de automação que existem no mercado e uma equipe de profissionais qualificados para atuar em diversos projetos, além de atender grandes demandas. ## Como se candidatar Enviar currículo com pretensão salarial para rh@xtesting.com.br
1.0
Analista de Qualidade na [XTESTING] - ### Analista de Qualidade Vaga para analista de qualidade na XTESTING ## Atividades Profissional responsável por atuar com autorias internas e externas, fazendo a elaboração de normas e procedimento e não conformidades dos processos de qualidade que se relaciona com todas as áreas ligadas a qualidade dentro de uma empresa. ## Local - Salvador ## Benefícios - Transporte; - Ticket Refeição; - Assistência médica; ## Requisitos **Obrigatórios:** - Nível superior completo; - Ter participado de projeto na área de qualidade; - Modelagem de processos; - Metodologias; - ISO 9001; **Diferenciais:** - MPT.Br; - CMMI; - MPS.Br; - SCRUM; ## Contratação - CLT; - Salário a combinar; ## XTESTING Fundada em 2013, a X-Testing é uma empresa especializada na prestação de serviços na área de teste de software com experiência na implantação de testes funcionais, de performance, usabilidade, segurança e acessibilidade em empresas privadas e do setor público. Promove a qualidade dos softwares desenvolvidos através de processos e metodologias certificados pelo modelo MPT. BR nível 4, utilizando as melhores ferramentas de automação que existem no mercado e uma equipe de profissionais qualificados para atuar em diversos projetos, além de atender grandes demandas. ## Como se candidatar Enviar currículo com pretensão salarial para rh@xtesting.com.br
process
analista de qualidade na analista de qualidade vaga para analista de qualidade na xtesting atividades profissional responsável por atuar com autorias internas e externas fazendo a elaboração de normas e procedimento e não conformidades dos processos de qualidade que se relaciona com todas as áreas ligadas a qualidade dentro de uma empresa local salvador benefícios transporte ticket refeição assistência médica requisitos obrigatórios nível superior completo ter participado de projeto na área de qualidade modelagem de processos metodologias iso diferenciais mpt br cmmi mps br scrum contratação clt salário a combinar xtesting fundada em a x testing é uma empresa especializada na prestação de serviços na área de teste de software com experiência na implantação de testes funcionais de performance usabilidade segurança e acessibilidade em empresas privadas e do setor público promove a qualidade dos softwares desenvolvidos através de processos e metodologias certificados pelo modelo mpt br nível utilizando as melhores ferramentas de automação que existem no mercado e uma equipe de profissionais qualificados para atuar em diversos projetos além de atender grandes demandas como se candidatar enviar currículo com pretensão salarial para rh xtesting com br
1
15,074
18,768,667,419
IssuesEvent
2021-11-06 12:11:36
google/android-fhir
https://api.github.com/repos/google/android-fhir
closed
Enable https://github.com/cashapp/licensee
Beta high priority process Q4 2021
**Describe the Issue** Enable the plugin and configure it so it allows only compatible licences in our dependencies. **Would you like to work on the issue?** @kevinmost made a headstart :)
1.0
Enable https://github.com/cashapp/licensee - **Describe the Issue** Enable the plugin and configure it so it allows only compatible licences in our dependencies. **Would you like to work on the issue?** @kevinmost made a headstart :)
process
enable describe the issue enable the plugin and configure it so it allows only compatible licences in our dependencies would you like to work on the issue kevinmost made a headstart
1
22,174
30,726,812,089
IssuesEvent
2023-07-27 20:24:25
computerise/stonks
https://api.github.com/repos/computerise/stonks
closed
Implement Weighted Average Cost of Capital for Discounted Cash Flow model
enhancement good first issue component:processing:model
Integrate computation of the [Weighted Average Cost of Capital (WACC)](https://www.investopedia.com/terms/w/wacc.asp) for use as the `discount_rate` parameter in the [Discounted Cash Flow (DCF)](https://www.investopedia.com/terms/d/dcf.asp) model.
1.0
Implement Weighted Average Cost of Capital for Discounted Cash Flow model - Integrate computation of the [Weighted Average Cost of Capital (WACC)](https://www.investopedia.com/terms/w/wacc.asp) for use as the `discount_rate` parameter in the [Discounted Cash Flow (DCF)](https://www.investopedia.com/terms/d/dcf.asp) model.
process
implement weighted average cost of capital for discounted cash flow model integrate computation of the for use as the discount rate parameter in the model
1
587,958
17,643,872,555
IssuesEvent
2021-08-20 01:06:37
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Namex API - Resubmit causes error in transaction history
bug NameX Priority1 ENTITY
**Describe the bug in current situation** When I resubmitted an NR and searched it in Namex, there was an error loading the transaction history. (GET events call received HTTP 500 - server error.) Comments from Severin: > This is needed unless the NR UI fix 8553 pre-empts it. > Also need to remove some code for Namex API (low priority). > No, it doesn’t fix the 500 in namex API when fetching events. @severinbeauvais can this wait until Thor et al cleans up the architecture in Names after NRO decommissioned? **Link bug to the User Story** **Impact of this bug** Describe the impact, i.e. what the impact is, and number of users impacted. **Chance of Occurring (high/medium/low/very low)** **Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?** Search NR 7312306 in Dev NameX **Steps to Reproduce** Steps to reproduce the behavior: 1. Resubmit an expired NR 2. Verify status on name request app (should be draft) 3. Go to NameX, verify status on NameX (should be draft) 4. Click on transaction history 5. See error loading ![image.png](https://images.zenhubusercontent.com/5d814bb8edbc510001817803/cdde0c5e-17d2-4f43-b26c-0c521c00fb0a) **Actual/ observed behavior/ results** **Expected behavior** GIVEN an expired NR has been resubmitted WHEN I search the NR in NameX THEN I should be able to retrieve a transaction history
1.0
Namex API - Resubmit causes error in transaction history - **Describe the bug in current situation** When I resubmitted an NR and searched it in Namex, there was an error loading the transaction history. (GET events call received HTTP 500 - server error.) Comments from Severin: > This is needed unless the NR UI fix 8553 pre-empts it. > Also need to remove some code for Namex API (low priority). > No, it doesn’t fix the 500 in namex API when fetching events. @severinbeauvais can this wait until Thor et al cleans up the architecture in Names after NRO decommissioned? **Link bug to the User Story** **Impact of this bug** Describe the impact, i.e. what the impact is, and number of users impacted. **Chance of Occurring (high/medium/low/very low)** **Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?** Search NR 7312306 in Dev NameX **Steps to Reproduce** Steps to reproduce the behavior: 1. Resubmit an expired NR 2. Verify status on name request app (should be draft) 3. Go to NameX, verify status on NameX (should be draft) 4. Click on transaction history 5. See error loading ![image.png](https://images.zenhubusercontent.com/5d814bb8edbc510001817803/cdde0c5e-17d2-4f43-b26c-0c521c00fb0a) **Actual/ observed behavior/ results** **Expected behavior** GIVEN an expired NR has been resubmitted WHEN I search the NR in NameX THEN I should be able to retrieve a transaction history
non_process
namex api resubmit causes error in transaction history describe the bug in current situation when i resubmitted an nr and searched it in namex there was an error loading the transaction history get events call received http server error comments from severin this is needed unless the nr ui fix pre empts it also need to remove some code for namex api low priority no it doesn’t fix the in namex api when fetching events severinbeauvais can this wait until thor et al cleans up the architecture in names after nro decommissioned link bug to the user story impact of this bug describe the impact i e what the impact is and number of users impacted chance of occurring high medium low very low pre conditions which env any pre requesites or assumptions to execute steps search nr in dev namex steps to reproduce steps to reproduce the behavior resubmit an expired nr verify status on name request app should be draft go to namex verify status on namex should be draft click on transaction history see error loading actual observed behavior results expected behavior given an expired nr has been resubmitted when i search the nr in namex then i should be able to retrieve a transaction history
0
15,427
19,618,630,842
IssuesEvent
2022-01-07 01:27:53
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
QgsProcessingAlgorithmDialogBase hides short help for ever
Processing Bug
### What is the bug or the crash? The QgsProcessingAlgorithmDialogBase is used to visualize parameters of a QgsProcessingAlgorithm. Calling `QgsProcessingAlgorithmDialogBase::setAlgorithm( QgsProcessingAlgorithm *algorithm )` changes the underlying algorithm and updates the dialog, e.g. the title and the algorithm help. If an algorithm does not provide a help string, the QTextBrowser to show this help is hidden. Although a new algorithm might provide a help string, the text browser will never become visible again. ![image](https://user-images.githubusercontent.com/1404870/148144590-d65f066e-6b36-4621-b81e-e1e9935252a4.png) ### Steps to reproduce the issue The following python code create two QgsProcessingAlgorithmDialogBase dialog instances (see screenshot above). Both are finally using the `native:rescaleraster` algorithm. In case of the HiddenHelp instance the help browser remains hidden, because the `gdal:aspect` algorithm, which does not provide a short help string, was used before. ````python from qgis.core import QgsApplication, QgsProcessingRegistry from qgis.gui import QgsProcessingAlgorithmDialogBase from qgis.testing.mocked import start_app APP = start_app() from processing.core.Processing import Processing from qgis.analysis import QgsNativeAlgorithms QgsApplication.processingRegistry().addProvider(QgsNativeAlgorithms()) Processing.initialize() reg: QgsProcessingRegistry = QgsApplication.instance().processingRegistry() alg1 = reg.algorithmById('native:rescaleraster') alg2 = reg.algorithmById('gdal:aspect') assert alg1.shortHelpString() != '' assert alg2.shortHelpString() == '' class ExampleDialog(QgsProcessingAlgorithmDialogBase): def __init__(self, *args, **kwds): super().__init__(*args, **kwds) VisibleHelp = ExampleDialog() VisibleHelp.setAlgorithm(alg1.create()) VisibleHelp.setWindowTitle(f'{VisibleHelp.windowTitle()} (Visible Help)') HiddenHelp = ExampleDialog() HiddenHelp.setAlgorithm(alg1.create()) HiddenHelp.setAlgorithm(alg2.create()) # no short help -> will hide the text browser HiddenHelp.setAlgorithm(alg1.create()) # text browser remains hidden HiddenHelp.setWindowTitle(f'{HiddenHelp.windowTitle()} (Hidden Help)') VisibleHelp.show() HiddenHelp.show() APP.exec_() ```` ### Versions QGIS version 3.23.0-Master QGIS code revision 307bb854b1 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
QgsProcessingAlgorithmDialogBase hides short help for ever - ### What is the bug or the crash? The QgsProcessingAlgorithmDialogBase is used to visualize parameters of a QgsProcessingAlgorithm. Calling `QgsProcessingAlgorithmDialogBase::setAlgorithm( QgsProcessingAlgorithm *algorithm )` changes the underlying algorithm and updates the dialog, e.g. the title and the algorithm help. If an algorithm does not provide a help string, the QTextBrowser to show this help is hidden. Although a new algorithm might provide a help string, the text browser will never become visible again. ![image](https://user-images.githubusercontent.com/1404870/148144590-d65f066e-6b36-4621-b81e-e1e9935252a4.png) ### Steps to reproduce the issue The following python code create two QgsProcessingAlgorithmDialogBase dialog instances (see screenshot above). Both are finally using the `native:rescaleraster` algorithm. In case of the HiddenHelp instance the help browser remains hidden, because the `gdal:aspect` algorithm, which does not provide a short help string, was used before. ````python from qgis.core import QgsApplication, QgsProcessingRegistry from qgis.gui import QgsProcessingAlgorithmDialogBase from qgis.testing.mocked import start_app APP = start_app() from processing.core.Processing import Processing from qgis.analysis import QgsNativeAlgorithms QgsApplication.processingRegistry().addProvider(QgsNativeAlgorithms()) Processing.initialize() reg: QgsProcessingRegistry = QgsApplication.instance().processingRegistry() alg1 = reg.algorithmById('native:rescaleraster') alg2 = reg.algorithmById('gdal:aspect') assert alg1.shortHelpString() != '' assert alg2.shortHelpString() == '' class ExampleDialog(QgsProcessingAlgorithmDialogBase): def __init__(self, *args, **kwds): super().__init__(*args, **kwds) VisibleHelp = ExampleDialog() VisibleHelp.setAlgorithm(alg1.create()) VisibleHelp.setWindowTitle(f'{VisibleHelp.windowTitle()} (Visible Help)') HiddenHelp = ExampleDialog() HiddenHelp.setAlgorithm(alg1.create()) HiddenHelp.setAlgorithm(alg2.create()) # no short help -> will hide the text browser HiddenHelp.setAlgorithm(alg1.create()) # text browser remains hidden HiddenHelp.setWindowTitle(f'{HiddenHelp.windowTitle()} (Hidden Help)') VisibleHelp.show() HiddenHelp.show() APP.exec_() ```` ### Versions QGIS version 3.23.0-Master QGIS code revision 307bb854b1 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
qgsprocessingalgorithmdialogbase hides short help for ever what is the bug or the crash the qgsprocessingalgorithmdialogbase is used to visualize parameters of a qgsprocessingalgorithm calling qgsprocessingalgorithmdialogbase setalgorithm qgsprocessingalgorithm algorithm changes the underlying algorithm and updates the dialog e g the title and the algorithm help if an algorithm does not provide a help string the qtextbrowser to show this help is hidden although a new algorithm might provide a help string the text browser will never become visible again steps to reproduce the issue the following python code create two qgsprocessingalgorithmdialogbase dialog instances see screenshot above both are finally using the native rescaleraster algorithm in case of the hiddenhelp instance the help browser remains hidden because the gdal aspect algorithm which does not provide a short help string was used before python from qgis core import qgsapplication qgsprocessingregistry from qgis gui import qgsprocessingalgorithmdialogbase from qgis testing mocked import start app app start app from processing core processing import processing from qgis analysis import qgsnativealgorithms qgsapplication processingregistry addprovider qgsnativealgorithms processing initialize reg qgsprocessingregistry qgsapplication instance processingregistry reg algorithmbyid native rescaleraster reg algorithmbyid gdal aspect assert shorthelpstring assert shorthelpstring class exampledialog qgsprocessingalgorithmdialogbase def init self args kwds super init args kwds visiblehelp exampledialog visiblehelp setalgorithm create visiblehelp setwindowtitle f visiblehelp windowtitle visible help hiddenhelp exampledialog hiddenhelp setalgorithm create hiddenhelp setalgorithm create no short help will hide the text browser hiddenhelp setalgorithm create text browser remains hidden hiddenhelp setwindowtitle f hiddenhelp windowtitle hidden help visiblehelp show hiddenhelp show app exec versions qgis version master qgis code revision supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
14,096
16,986,220,681
IssuesEvent
2021-06-30 14:39:54
audacity/audacity
https://api.github.com/repos/audacity/audacity
reopened
QA: Make a bug review in Bugzilla and Github
QA process
Check relevance. Are all bugs reproducible with a current stable release? Check and correct bug priority ( if necessary)
1.0
QA: Make a bug review in Bugzilla and Github - Check relevance. Are all bugs reproducible with a current stable release? Check and correct bug priority ( if necessary)
process
qa make a bug review in bugzilla and github check relevance are all bugs reproducible with a current stable release check and correct bug priority if necessary
1
5,420
8,257,026,395
IssuesEvent
2018-09-13 02:28:27
uccser/verto
https://api.github.com/repos/uccser/verto
opened
Add blockquote tag
processor implementation
The standard Markdown blockquote tag is very limited in formatting, and it would be nice to have a Verto tag to allow editing, especially for use with [Bootstrap](https://getbootstrap.com/docs/4.1/content/typography/#blockquotes). Possibly could look like: ```markdown {blockquote} First and foremost, we believe that speed is more than a feature. - [Fred Wilson](https://en.wikipedia.org/wiki/Fred_Wilson_(financier)) {blockquote end} ``` This could be used with a template like the following: ```html <blockquote class="blockquote"> {{ content }} <footer class="blockquote-footer"> {{ source_content}} </footer> </blockquote> ``` to produce something like the following ```html <blockquote class="blockquote"> <p>First and foremost, we believe that speed is more than a feature.</p> <footer class="blockquote-footer"> [Fred Wilson](https://en.wikipedia.org/wiki/Fred_Wilson_(financier)) </footer> </blockquote> ``` Would need to figure out how to detect footer information (could be argument value like image `alt` text, though it wouldn't work for in-context translation (though this may be a worth compromise for a simpler tag). ## Optional arguments - **align** - Text value to be passed through to template. - **source** - URL text value to be passed through to the template.
1.0
Add blockquote tag - The standard Markdown blockquote tag is very limited in formatting, and it would be nice to have a Verto tag to allow editing, especially for use with [Bootstrap](https://getbootstrap.com/docs/4.1/content/typography/#blockquotes). Possibly could look like: ```markdown {blockquote} First and foremost, we believe that speed is more than a feature. - [Fred Wilson](https://en.wikipedia.org/wiki/Fred_Wilson_(financier)) {blockquote end} ``` This could be used with a template like the following: ```html <blockquote class="blockquote"> {{ content }} <footer class="blockquote-footer"> {{ source_content}} </footer> </blockquote> ``` to produce something like the following ```html <blockquote class="blockquote"> <p>First and foremost, we believe that speed is more than a feature.</p> <footer class="blockquote-footer"> [Fred Wilson](https://en.wikipedia.org/wiki/Fred_Wilson_(financier)) </footer> </blockquote> ``` Would need to figure out how to detect footer information (could be argument value like image `alt` text, though it wouldn't work for in-context translation (though this may be a worth compromise for a simpler tag). ## Optional arguments - **align** - Text value to be passed through to template. - **source** - URL text value to be passed through to the template.
process
add blockquote tag the standard markdown blockquote tag is very limited in formatting and it would be nice to have a verto tag to allow editing especially for use with possibly could look like markdown blockquote first and foremost we believe that speed is more than a feature blockquote end this could be used with a template like the following html content source content to produce something like the following html first and foremost we believe that speed is more than a feature would need to figure out how to detect footer information could be argument value like image alt text though it wouldn t work for in context translation though this may be a worth compromise for a simpler tag optional arguments align text value to be passed through to template source url text value to be passed through to the template
1
20,003
26,478,932,488
IssuesEvent
2023-01-17 13:20:35
googleapis/ruby-spanner-activerecord
https://api.github.com/repos/googleapis/ruby-spanner-activerecord
closed
ActiveRecord::Model::LoggingTest#test_logs_without_binds is flaky
api: spanner type: process priority: p2
``` 1) Failure: ActiveRecord::Model::LoggingTest#test_logs_without_binds [/home/runner/work/ruby-spanner-activerecord/ruby-spanner-activerecord/acceptance/cases/models/logging_test.rb:30]: Expected: 1 Actual: 0 2) Failure: ActiveRecord::Model::LoggingTest#test_logs_with_binds [/home/runner/work/ruby-spanner-activerecord/ruby-spanner-activerecord/acceptance/cases/models/logging_test.rb:41]: Expected: 1 ```
1.0
ActiveRecord::Model::LoggingTest#test_logs_without_binds is flaky - ``` 1) Failure: ActiveRecord::Model::LoggingTest#test_logs_without_binds [/home/runner/work/ruby-spanner-activerecord/ruby-spanner-activerecord/acceptance/cases/models/logging_test.rb:30]: Expected: 1 Actual: 0 2) Failure: ActiveRecord::Model::LoggingTest#test_logs_with_binds [/home/runner/work/ruby-spanner-activerecord/ruby-spanner-activerecord/acceptance/cases/models/logging_test.rb:41]: Expected: 1 ```
process
activerecord model loggingtest test logs without binds is flaky failure activerecord model loggingtest test logs without binds expected actual failure activerecord model loggingtest test logs with binds expected
1
21,192
28,209,493,761
IssuesEvent
2023-04-05 02:02:33
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE?] rt.com
whitelisting process
**Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. `www.rt.com` **More Information** <!-- How did you discover your web site or domain was listed here? --> 1. tried to go to the website, it doesn't belong to me. **Have you requested removal from other sources?** <!-- Please include all relevant links to your existing removals / whitelistings. --> none ... **Additional context** <!-- Add any other context about the problem here. --> I copied your host file into my host file under the assumption that this list is not a political censorship list and not some kind of "truth" list. If this is not the case but it's also a political list and truth list, then that's perfectly fine too, but please make it clear that ideology and censorship is one of the purposes of this block list. <!-- ❗ We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process. -->
1.0
[FALSE-POSITIVE?] rt.com - **Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. `www.rt.com` **More Information** <!-- How did you discover your web site or domain was listed here? --> 1. tried to go to the website, it doesn't belong to me. **Have you requested removal from other sources?** <!-- Please include all relevant links to your existing removals / whitelistings. --> none ... **Additional context** <!-- Add any other context about the problem here. --> I copied your host file into my host file under the assumption that this list is not a political censorship list and not some kind of "truth" list. If this is not the case but it's also a political list and truth list, then that's perfectly fine too, but please make it clear that ideology and censorship is one of the purposes of this block list. <!-- ❗ We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process. -->
process
rt com domains or links more information tried to go to the website it doesn t belong to me have you requested removal from other sources none additional context i copied your host file into my host file under the assumption that this list is not a political censorship list and not some kind of truth list if this is not the case but it s also a political list and truth list then that s perfectly fine too but please make it clear that ideology and censorship is one of the purposes of this block list ❗ we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
1
18,896
24,835,520,967
IssuesEvent
2022-10-26 08:32:33
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
opened
[General] Darstellung Gebäude
webform delete-after-processing general
Liebes App Team, zunächst einmal erfreulich, dass es es endlich interaktive Karten für die Lage von TUM-Räumen gibt. Bezüglich der Darstellung möchte ich allerdings anregen Gebäude nicht in der Farbe Blau darzustellen. Die TUM-Farbe in allen Ehren, die Farbe symbolisiert auf karten in der Regel Gewässer und der Lesbarkeit nicht dienlich. Schöne Grüße, Sebastian Haß -- Dipl. Ing. Sebastian Haß Officer of the Vice Dean Academic and Student Affairs TUM School of Engineering and Design
1.0
[General] Darstellung Gebäude - Liebes App Team, zunächst einmal erfreulich, dass es es endlich interaktive Karten für die Lage von TUM-Räumen gibt. Bezüglich der Darstellung möchte ich allerdings anregen Gebäude nicht in der Farbe Blau darzustellen. Die TUM-Farbe in allen Ehren, die Farbe symbolisiert auf karten in der Regel Gewässer und der Lesbarkeit nicht dienlich. Schöne Grüße, Sebastian Haß -- Dipl. Ing. Sebastian Haß Officer of the Vice Dean Academic and Student Affairs TUM School of Engineering and Design
process
darstellung gebäude liebes app team zunächst einmal erfreulich dass es es endlich interaktive karten für die lage von tum räumen gibt bezüglich der darstellung möchte ich allerdings anregen gebäude nicht in der farbe blau darzustellen die tum farbe in allen ehren die farbe symbolisiert auf karten in der regel gewässer und der lesbarkeit nicht dienlich schöne grüße sebastian haß dipl ing sebastian haß officer of the vice dean academic and student affairs tum school of engineering and design
1
20,762
27,494,147,381
IssuesEvent
2023-03-05 00:30:54
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
opened
Create an account recovery system backend and integrate it with the frontend.
Processing Task Sprint 2
**Task Test* *Test 1* 1. Go to the account recovery page 2. Enter email and click submit 3. Check if password returned is correct *Test 2* 1. Go to the account recovery page 2. Enter email not in database and click submit 3. Check if it returns a password.
1.0
Create an account recovery system backend and integrate it with the frontend. - **Task Test* *Test 1* 1. Go to the account recovery page 2. Enter email and click submit 3. Check if password returned is correct *Test 2* 1. Go to the account recovery page 2. Enter email not in database and click submit 3. Check if it returns a password.
process
create an account recovery system backend and integrate it with the frontend task test test go to the account recovery page enter email and click submit check if password returned is correct test go to the account recovery page enter email not in database and click submit check if it returns a password
1
117,216
15,077,754,385
IssuesEvent
2021-02-05 07:33:38
wso2/cellery
https://api.github.com/repos/wso2/cellery
closed
Cellery IDE should allow linking cells and adding environment variables
Resolution/Won’t Fix component/designer
**Description:** Currently (0.6.0) Cellery IDE is not supporting linking cells and adding environment variables. This should be implemented in a future release. **Suggested Labels:** New feature
1.0
Cellery IDE should allow linking cells and adding environment variables - **Description:** Currently (0.6.0) Cellery IDE is not supporting linking cells and adding environment variables. This should be implemented in a future release. **Suggested Labels:** New feature
non_process
cellery ide should allow linking cells and adding environment variables description currently cellery ide is not supporting linking cells and adding environment variables this should be implemented in a future release suggested labels new feature
0
10,207
14,876,711,944
IssuesEvent
2021-01-20 01:28:35
DualSaturn/wi21-cse110-lab3
https://api.github.com/repos/DualSaturn/wi21-cse110-lab3
closed
Need all the selector grouping/combinators to be used
requirement
Selector List (element, element) Descendant Combinator (element element) Child Combinator (element > element) General sibling combinator (element + element) Adjacent sibling combinator (element ~ element) Combining Two Selectors (element.class)
1.0
Need all the selector grouping/combinators to be used - Selector List (element, element) Descendant Combinator (element element) Child Combinator (element > element) General sibling combinator (element + element) Adjacent sibling combinator (element ~ element) Combining Two Selectors (element.class)
non_process
need all the selector grouping combinators to be used selector list element element descendant combinator element element child combinator element element general sibling combinator element element adjacent sibling combinator element element combining two selectors element class
0
106,374
16,673,330,283
IssuesEvent
2021-06-07 13:34:19
VivekBuzruk/Hygieia
https://api.github.com/repos/VivekBuzruk/Hygieia
closed
WS-2020-0218 (High) detected in merge-1.2.1.tgz - autoclosed
security vulnerability
## WS-2020-0218 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: Hygieia/UI/package.json</p> <p>Path to vulnerable library: Hygieia/UI/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - gulp-ng-annotate-0.5.3.tgz (Root Library) - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/VivekBuzruk/Hygieia/commit/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Prototype Pollution vulnerability was found in merge before 2.1.0 via the merge.recursive function. It can be tricked into adding or modifying properties of the Object prototype. These properties will be present on all objects. <p>Publish Date: 2020-10-09 <p>URL: <a href=https://github.com/yeikos/js.merge/pull/38>WS-2020-0218</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p> <p>Release Date: 2020-10-09</p> <p>Fix Resolution: merge - 2.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0218 (High) detected in merge-1.2.1.tgz - autoclosed - ## WS-2020-0218 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: Hygieia/UI/package.json</p> <p>Path to vulnerable library: Hygieia/UI/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - gulp-ng-annotate-0.5.3.tgz (Root Library) - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/VivekBuzruk/Hygieia/commit/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Prototype Pollution vulnerability was found in merge before 2.1.0 via the merge.recursive function. It can be tricked into adding or modifying properties of the Object prototype. These properties will be present on all objects. <p>Publish Date: 2020-10-09 <p>URL: <a href=https://github.com/yeikos/js.merge/pull/38>WS-2020-0218</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p> <p>Release Date: 2020-10-09</p> <p>Fix Resolution: merge - 2.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in merge tgz autoclosed ws high severity vulnerability vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file hygieia ui package json path to vulnerable library hygieia ui node modules merge package json dependency hierarchy gulp ng annotate tgz root library x merge tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in merge before via the merge recursive function it can be tricked into adding or modifying properties of the object prototype these properties will be present on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution merge step up your open source security game with whitesource
0
12,758
15,114,322,004
IssuesEvent
2021-02-09 01:38:56
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
How to analyze zipped log files on a remote webserver?
log-processing question
I want to set up goaccess on my local RaspberryPi and after that I want to analyze logfiles from a few remote server of the same provider. This provider stores all the logfiles in /logs So far so good... But... all these logfiles are zipped like: access_log_2020-03-03.gz Is it somehow possible to read these zipped logfiles with goaccess? Thanks in advance... :)
1.0
How to analyze zipped log files on a remote webserver? - I want to set up goaccess on my local RaspberryPi and after that I want to analyze logfiles from a few remote server of the same provider. This provider stores all the logfiles in /logs So far so good... But... all these logfiles are zipped like: access_log_2020-03-03.gz Is it somehow possible to read these zipped logfiles with goaccess? Thanks in advance... :)
process
how to analyze zipped log files on a remote webserver i want to set up goaccess on my local raspberrypi and after that i want to analyze logfiles from a few remote server of the same provider this provider stores all the logfiles in logs so far so good but all these logfiles are zipped like access log gz is it somehow possible to read these zipped logfiles with goaccess thanks in advance
1
21,374
3,701,734,057
IssuesEvent
2016-02-29 14:28:34
mysociety/pombola
https://api.github.com/repos/mysociety/pombola
closed
Import constituency office data and photos
4 - Reviewing data-import design Difficulty 2 Rep Locator ZA
PMG have been undertaking an exercise to collect accurate locations and photos of all the party constituency offices. We need: - [x] A new script to import their data and update our database with the correct locations - [x] The script to also store copies of the photos - [x] To display the photos on each constituency office page [do #1948 at same time]
1.0
Import constituency office data and photos - PMG have been undertaking an exercise to collect accurate locations and photos of all the party constituency offices. We need: - [x] A new script to import their data and update our database with the correct locations - [x] The script to also store copies of the photos - [x] To display the photos on each constituency office page [do #1948 at same time]
non_process
import constituency office data and photos pmg have been undertaking an exercise to collect accurate locations and photos of all the party constituency offices we need a new script to import their data and update our database with the correct locations the script to also store copies of the photos to display the photos on each constituency office page
0
17,370
23,192,649,789
IssuesEvent
2022-08-01 13:53:24
vacp2p/research
https://api.github.com/repos/vacp2p/research
opened
[WIP] Vac RFC Process Roadmap
track:rfc-process
This roadmap tracks milestones and issues related to the Vac RFC process. ## Milestones and Issues * [ ] Vac RFC categories & structure * [ ] Update existing RFCs to new structure (on status promotion)
1.0
[WIP] Vac RFC Process Roadmap - This roadmap tracks milestones and issues related to the Vac RFC process. ## Milestones and Issues * [ ] Vac RFC categories & structure * [ ] Update existing RFCs to new structure (on status promotion)
process
vac rfc process roadmap this roadmap tracks milestones and issues related to the vac rfc process milestones and issues vac rfc categories structure update existing rfcs to new structure on status promotion
1
16,734
2,941,305,416
IssuesEvent
2015-07-02 06:49:26
tnt944445/reaver-wps
https://api.github.com/repos/tnt944445/reaver-wps
closed
Source code patch to correct the skipping of pin checks
auto-migrated Priority-Triage Type-Defect
``` As described in issue #461 the bug caused skipping pin tests when checking the second portion of the pin. Let me introduce this simple patch here. You have to have a source files tree of reaver (at best the svn checkout) and complete gcc toolchain installed and working to compile the sources. Locate the file ../src/exchange.c and somewhere around the line # 233 find this text else if(premature_timeout) { /* * Some WPS implementations simply drop the connection on the floor instead of sending a NACK. * We need to be able to handle this, but at the same time using a timeout on the M5/M7 messages * can result in false negatives. Thus, treating M5/M7 receive timeouts as NACKs can be disabled. * Only treat the timeout as a NACK if this feature is enabled. */ if(get_timeout_is_nack() && (last_msg == M3 || last_msg == M5) and in text editor replace the last condition line with the line ((last_msg == M3 && (get_key_status() == KEY1_WIP)) || last_msg == M5)) This condition checks additionally the status of the state machine and type of the last received message. So the condition is true only when we are in status KEY1_WIP [ie. trying the first portion of the pin] AND after receiving M3 OR just receiving M5. Without this check it returned the wrong ret_val = KEY_REJECTED and skipped the actual pin re-try. For completness, the same condition a few lines above of the previous if(got_nack) block should be written accordingly although I'm not sure if such condition state could occurr. Tested and now it works as expected. My setup is Debian Wheezy, rt73usb/zd1211rw, router Tenda W311, reaver 1.4 souces svn rev 113. Hope it helps. Please test and respond with results. ``` Original issue reported on code.google.com by `agent...@gmail.com` on 14 Feb 2013 at 4:34
1.0
Source code patch to correct the skipping of pin checks - ``` As described in issue #461 the bug caused skipping pin tests when checking the second portion of the pin. Let me introduce this simple patch here. You have to have a source files tree of reaver (at best the svn checkout) and complete gcc toolchain installed and working to compile the sources. Locate the file ../src/exchange.c and somewhere around the line # 233 find this text else if(premature_timeout) { /* * Some WPS implementations simply drop the connection on the floor instead of sending a NACK. * We need to be able to handle this, but at the same time using a timeout on the M5/M7 messages * can result in false negatives. Thus, treating M5/M7 receive timeouts as NACKs can be disabled. * Only treat the timeout as a NACK if this feature is enabled. */ if(get_timeout_is_nack() && (last_msg == M3 || last_msg == M5) and in text editor replace the last condition line with the line ((last_msg == M3 && (get_key_status() == KEY1_WIP)) || last_msg == M5)) This condition checks additionally the status of the state machine and type of the last received message. So the condition is true only when we are in status KEY1_WIP [ie. trying the first portion of the pin] AND after receiving M3 OR just receiving M5. Without this check it returned the wrong ret_val = KEY_REJECTED and skipped the actual pin re-try. For completness, the same condition a few lines above of the previous if(got_nack) block should be written accordingly although I'm not sure if such condition state could occurr. Tested and now it works as expected. My setup is Debian Wheezy, rt73usb/zd1211rw, router Tenda W311, reaver 1.4 souces svn rev 113. Hope it helps. Please test and respond with results. ``` Original issue reported on code.google.com by `agent...@gmail.com` on 14 Feb 2013 at 4:34
non_process
source code patch to correct the skipping of pin checks as described in issue the bug caused skipping pin tests when checking the second portion of the pin let me introduce this simple patch here you have to have a source files tree of reaver at best the svn checkout and complete gcc toolchain installed and working to compile the sources locate the file src exchange c and somewhere around the line find this text else if premature timeout some wps implementations simply drop the connection on the floor instead of sending a nack we need to be able to handle this but at the same time using a timeout on the messages can result in false negatives thus treating receive timeouts as nacks can be disabled only treat the timeout as a nack if this feature is enabled if get timeout is nack last msg last msg and in text editor replace the last condition line with the line last msg get key status wip last msg this condition checks additionally the status of the state machine and type of the last received message so the condition is true only when we are in status wip and after receiving or just receiving without this check it returned the wrong ret val key rejected and skipped the actual pin re try for completness the same condition a few lines above of the previous if got nack block should be written accordingly although i m not sure if such condition state could occurr tested and now it works as expected my setup is debian wheezy router tenda reaver souces svn rev hope it helps please test and respond with results original issue reported on code google com by agent gmail com on feb at
0
601,839
18,436,551,987
IssuesEvent
2021-10-14 13:37:22
airshipit/airshipctl
https://api.github.com/repos/airshipit/airshipctl
closed
Upgrade CAPI to v1alpha4 & CAPM3 to v1alpha5 for Bare Metal
enhancement priority/critical 6-upstream/metal3-io size l
Once it is been released for GA, upgrade CAPI components from v0.3.x to v0.4.x (aka v1alpha4). There will also be a need to uplift a corresponding version of CAPM3 from v0.4.2 to v0.5.0 (aka v1alpha5) at the same time. When upgrading, the following need to be taken into consideration: - Are there breaking changes that require coding/configuration changes for the CAPI components? - Are there bug fixes in the upgrades that fix issues encountered when deploying? If so, are there workarounds in place today that would need to be removed or deprecated? May need input from @sb464f and others who have been running deployments. - Are there any new features in the upgrades that address missing capabilities which have hindered the deployments? In utilizing these new features, what workarounds are in place today that would need to be removed or deprecated? May need input from @sb464f and others who have been running deployments. Sources Cluster API - https://cluster-api.sigs.k8s.io/developer/providers/v1alpha3-to-v1alpha4.html - https://github.com/kubernetes-sigs/cluster-api CAPM3 - https://github.com/metal3-io/cluster-api-provider-metal3 The goal of this issue is to upgrade the CAPI & CAPM3 components, but leave current functionality in place as much as possible. If there are new features in the upgrade that Airshipctl can take advantage of, these should be reviewed on the Design Call & new issues created to implement. This should occur after the CAPM3, BMO & Ironic upgrades being performed in #554
1.0
Upgrade CAPI to v1alpha4 & CAPM3 to v1alpha5 for Bare Metal - Once it is been released for GA, upgrade CAPI components from v0.3.x to v0.4.x (aka v1alpha4). There will also be a need to uplift a corresponding version of CAPM3 from v0.4.2 to v0.5.0 (aka v1alpha5) at the same time. When upgrading, the following need to be taken into consideration: - Are there breaking changes that require coding/configuration changes for the CAPI components? - Are there bug fixes in the upgrades that fix issues encountered when deploying? If so, are there workarounds in place today that would need to be removed or deprecated? May need input from @sb464f and others who have been running deployments. - Are there any new features in the upgrades that address missing capabilities which have hindered the deployments? In utilizing these new features, what workarounds are in place today that would need to be removed or deprecated? May need input from @sb464f and others who have been running deployments. Sources Cluster API - https://cluster-api.sigs.k8s.io/developer/providers/v1alpha3-to-v1alpha4.html - https://github.com/kubernetes-sigs/cluster-api CAPM3 - https://github.com/metal3-io/cluster-api-provider-metal3 The goal of this issue is to upgrade the CAPI & CAPM3 components, but leave current functionality in place as much as possible. If there are new features in the upgrade that Airshipctl can take advantage of, these should be reviewed on the Design Call & new issues created to implement. This should occur after the CAPM3, BMO & Ironic upgrades being performed in #554
non_process
upgrade capi to to for bare metal once it is been released for ga upgrade capi components from x to x aka there will also be a need to uplift a corresponding version of from to aka at the same time when upgrading the following need to be taken into consideration are there breaking changes that require coding configuration changes for the capi components are there bug fixes in the upgrades that fix issues encountered when deploying if so are there workarounds in place today that would need to be removed or deprecated may need input from and others who have been running deployments are there any new features in the upgrades that address missing capabilities which have hindered the deployments in utilizing these new features what workarounds are in place today that would need to be removed or deprecated may need input from and others who have been running deployments sources cluster api the goal of this issue is to upgrade the capi components but leave current functionality in place as much as possible if there are new features in the upgrade that airshipctl can take advantage of these should be reviewed on the design call new issues created to implement this should occur after the bmo ironic upgrades being performed in
0
184,423
31,896,481,126
IssuesEvent
2023-09-18 02:32:33
agency-of-learning/PairApp
https://api.github.com/repos/agency-of-learning/PairApp
closed
[UserMenteeApplication] We should be notified when a new application is submitted
design doc required high priority epic
When someone submits an application, we should be notified by email that an application has been submitted. We should also send an email to the user telling them that we have received their application and linking them to the show page (user mentee application). We should also send a notification to the admin to let them know a new application has been submitted.
1.0
[UserMenteeApplication] We should be notified when a new application is submitted - When someone submits an application, we should be notified by email that an application has been submitted. We should also send an email to the user telling them that we have received their application and linking them to the show page (user mentee application). We should also send a notification to the admin to let them know a new application has been submitted.
non_process
we should be notified when a new application is submitted when someone submits an application we should be notified by email that an application has been submitted we should also send an email to the user telling them that we have received their application and linking them to the show page user mentee application we should also send a notification to the admin to let them know a new application has been submitted
0
3,656
6,691,854,893
IssuesEvent
2017-10-09 14:30:35
oasis-tcs/sarif-spec
https://api.github.com/repos/oasis-tcs/sarif-spec
closed
Define and document issue workflow
process
A preliminary proposal for driving issue workflow for spec changes. We should refine, add to and subtract from this content in discussion here. - All activity that may impact spec content will be tracked as issues, e.g., https://github.com/oasis-tcs/sarif-spec/issues/1 - We encourage discussion on the public issues. Editors will ‘curate’ by applying specific tags, ensuring content from other sources (email) is recorded, etc. - Editors will prepare proposed changes in separate repository branches. A clean document with revisions tracking enabled will be used for driving specific textual proposals. - Any issues that require time-sensitive discussion will be driven through the mailing list (this should be rare) - Issues that are ready for final approval or which warrant discussion on the telecon will be tagged in advance of each meeting. - After telecon approval/rejection, PRs will be merged or closed unmerged as appropriate. - This process (with additional details) will be driven by an open issue and documented in a file persisted to the repository (and will be approved next TC)
1.0
Define and document issue workflow - A preliminary proposal for driving issue workflow for spec changes. We should refine, add to and subtract from this content in discussion here. - All activity that may impact spec content will be tracked as issues, e.g., https://github.com/oasis-tcs/sarif-spec/issues/1 - We encourage discussion on the public issues. Editors will ‘curate’ by applying specific tags, ensuring content from other sources (email) is recorded, etc. - Editors will prepare proposed changes in separate repository branches. A clean document with revisions tracking enabled will be used for driving specific textual proposals. - Any issues that require time-sensitive discussion will be driven through the mailing list (this should be rare) - Issues that are ready for final approval or which warrant discussion on the telecon will be tagged in advance of each meeting. - After telecon approval/rejection, PRs will be merged or closed unmerged as appropriate. - This process (with additional details) will be driven by an open issue and documented in a file persisted to the repository (and will be approved next TC)
process
define and document issue workflow a preliminary proposal for driving issue workflow for spec changes we should refine add to and subtract from this content in discussion here all activity that may impact spec content will be tracked as issues e g we encourage discussion on the public issues editors will ‘curate’ by applying specific tags ensuring content from other sources email is recorded etc editors will prepare proposed changes in separate repository branches a clean document with revisions tracking enabled will be used for driving specific textual proposals any issues that require time sensitive discussion will be driven through the mailing list this should be rare issues that are ready for final approval or which warrant discussion on the telecon will be tagged in advance of each meeting after telecon approval rejection prs will be merged or closed unmerged as appropriate this process with additional details will be driven by an open issue and documented in a file persisted to the repository and will be approved next tc
1
308,132
23,233,766,552
IssuesEvent
2022-08-03 09:53:58
ShawnLemuelDabi/ITP-Group-18
https://api.github.com/repos/ShawnLemuelDabi/ITP-Group-18
opened
[DOCUMENTATION] - Document User Feedback from future NTUC Workshops
documentation
- [ ] Document User feedback from Workshop 2 - [ ] Document User feedback from Workshop 3 - [ ] Document User feedback from Workshop 4
1.0
[DOCUMENTATION] - Document User Feedback from future NTUC Workshops - - [ ] Document User feedback from Workshop 2 - [ ] Document User feedback from Workshop 3 - [ ] Document User feedback from Workshop 4
non_process
document user feedback from future ntuc workshops document user feedback from workshop document user feedback from workshop document user feedback from workshop
0
5,439
8,303,330,733
IssuesEvent
2018-09-21 17:11:28
nodejs/node
https://api.github.com/repos/nodejs/node
reopened
Setting callback for on 'data' event of stderr causes error on v10.0.0
confirmed-bug net process stream
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: 10.0.0 * **Platform**: OS X 10.13.2 * **Subsystem**: Process <!-- Enter your issue details below this comment. --> I am unable to set a callback on the 'data' event (since `process.stderr` [is a](https://nodejs.org/api/process.html#process_process_stdin) `net.Socket` stream, [it should](https://nodejs.org/api/stream.html#stream_event_data) emit the `data` event). The following snippet works on v8.0.0 but throws an error on v10.0.0: ```javascript process.stderr.on('data', function(data) { // not relevant, error is threw anyway }); ``` Error threw: ``` Error: read ENOTCONN at WriteStream.Socket._read (net.js:530:20) at WriteStream.Readable.read (_stream_readable.js:458:10) at resume_ (_stream_readable.js:897:12) at process._tickCallback (internal/process/next_tick.js:174:19) at Function.Module.runMain (internal/modules/cjs/loader.js:721:11) at startup (internal/bootstrap/node.js:228:19) at bootstrapNodeJSCore (internal/bootstrap/node.js:575:3) Emitted 'error' event at: at emitErrorNT (internal/streams/destroy.js:92:8) at emitErrorAndCloseNT (internal/streams/destroy.js:59:3) at process._tickCallback (internal/process/next_tick.js:174:19) [... lines matching original stack trace ...] at bootstrapNodeJSCore (internal/bootstrap/node.js:575:3) ````
1.0
Setting callback for on 'data' event of stderr causes error on v10.0.0 - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: 10.0.0 * **Platform**: OS X 10.13.2 * **Subsystem**: Process <!-- Enter your issue details below this comment. --> I am unable to set a callback on the 'data' event (since `process.stderr` [is a](https://nodejs.org/api/process.html#process_process_stdin) `net.Socket` stream, [it should](https://nodejs.org/api/stream.html#stream_event_data) emit the `data` event). The following snippet works on v8.0.0 but throws an error on v10.0.0: ```javascript process.stderr.on('data', function(data) { // not relevant, error is threw anyway }); ``` Error threw: ``` Error: read ENOTCONN at WriteStream.Socket._read (net.js:530:20) at WriteStream.Readable.read (_stream_readable.js:458:10) at resume_ (_stream_readable.js:897:12) at process._tickCallback (internal/process/next_tick.js:174:19) at Function.Module.runMain (internal/modules/cjs/loader.js:721:11) at startup (internal/bootstrap/node.js:228:19) at bootstrapNodeJSCore (internal/bootstrap/node.js:575:3) Emitted 'error' event at: at emitErrorNT (internal/streams/destroy.js:92:8) at emitErrorAndCloseNT (internal/streams/destroy.js:59:3) at process._tickCallback (internal/process/next_tick.js:174:19) [... lines matching original stack trace ...] at bootstrapNodeJSCore (internal/bootstrap/node.js:575:3) ````
process
setting callback for on data event of stderr causes error on thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform os x subsystem process i am unable to set a callback on the data event since process stderr net socket stream emit the data event the following snippet works on but throws an error on javascript process stderr on data function data not relevant error is threw anyway error threw error read enotconn at writestream socket read net js at writestream readable read stream readable js at resume stream readable js at process tickcallback internal process next tick js at function module runmain internal modules cjs loader js at startup internal bootstrap node js at bootstrapnodejscore internal bootstrap node js emitted error event at at emiterrornt internal streams destroy js at emiterrorandclosent internal streams destroy js at process tickcallback internal process next tick js at bootstrapnodejscore internal bootstrap node js
1
191,230
6,827,068,152
IssuesEvent
2017-11-08 15:58:09
b3aver/Automate
https://api.github.com/repos/b3aver/Automate
closed
In viewMode show Actions without the input fields
priority:minor time:medium topic:ui type:enhancement
- [x] Insert span tags with the saved informations. Substitute also the select fields. - [x] shorten urls and selectors on click show the entire field
1.0
In viewMode show Actions without the input fields - - [x] Insert span tags with the saved informations. Substitute also the select fields. - [x] shorten urls and selectors on click show the entire field
non_process
in viewmode show actions without the input fields insert span tags with the saved informations substitute also the select fields shorten urls and selectors on click show the entire field
0
449,141
31,829,580,731
IssuesEvent
2023-09-14 09:42:34
celestiaorg/celestia-app
https://api.github.com/repos/celestiaorg/celestia-app
closed
mint.BondDenom is not changeable via governance
documentation specs
## Context https://github.com/celestiaorg/celestia-app/blob/41ead770b48d6b86dce3f252e1be844b4978e907/specs/src/specs/params.md?plain=1#L55 ## Problem BondDenom is a genesis param and isn't governance modifiable ## Proposal Update the boolean in the last column to state that this param isn't governance modifiable
1.0
mint.BondDenom is not changeable via governance - ## Context https://github.com/celestiaorg/celestia-app/blob/41ead770b48d6b86dce3f252e1be844b4978e907/specs/src/specs/params.md?plain=1#L55 ## Problem BondDenom is a genesis param and isn't governance modifiable ## Proposal Update the boolean in the last column to state that this param isn't governance modifiable
non_process
mint bonddenom is not changeable via governance context problem bonddenom is a genesis param and isn t governance modifiable proposal update the boolean in the last column to state that this param isn t governance modifiable
0
21,263
28,438,588,274
IssuesEvent
2023-04-15 16:17:03
sulton-max/profile.todoapp
https://api.github.com/repos/sulton-max/profile.todoapp
opened
Create Entity Processing Base Service
processing
# The Ask Develop Entity Processing Base Service with generic repository pattern # How to Complete this Task Here's some steps to complete this deliverable. - Create contract for processing service - Create implementation
1.0
Create Entity Processing Base Service - # The Ask Develop Entity Processing Base Service with generic repository pattern # How to Complete this Task Here's some steps to complete this deliverable. - Create contract for processing service - Create implementation
process
create entity processing base service the ask develop entity processing base service with generic repository pattern how to complete this task here s some steps to complete this deliverable create contract for processing service create implementation
1
74,243
14,224,120,365
IssuesEvent
2020-11-17 19:11:40
foss-responders/fossresponders.com
https://api.github.com/repos/foss-responders/fossresponders.com
closed
Add jekyll-assets plugin
code
Right now, the Privacy page was removed partially because the asset path changes for images on secondary pages. We need an asset manager. Jekyll-asset should work. This will be important for the blog, too.
1.0
Add jekyll-assets plugin - Right now, the Privacy page was removed partially because the asset path changes for images on secondary pages. We need an asset manager. Jekyll-asset should work. This will be important for the blog, too.
non_process
add jekyll assets plugin right now the privacy page was removed partially because the asset path changes for images on secondary pages we need an asset manager jekyll asset should work this will be important for the blog too
0
168,441
20,764,170,046
IssuesEvent
2022-03-15 18:59:25
husnuljahneer/PortfolioReact
https://api.github.com/repos/husnuljahneer/PortfolioReact
opened
eslint-7.17.0.tgz: 1 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eslint-7.17.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /port/functions/package.json</p> <p>Path to vulnerable library: /functions/node_modules/ansi-regex/package.json,/port/node_modules/ansi-regex/package.json,/port/functions/node_modules/ansi-regex/package.json,/node_modules/ansi-regex/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/husnuljahneer/PortfolioReact/commit/965058239ea8a371ffb53aab2e372f73b551779f">965058239ea8a371ffb53aab2e372f73b551779f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-3807](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | ansi-regex-5.0.0.tgz | Transitive | 7.18.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3807</summary> ### Vulnerable Library - <b>ansi-regex-5.0.0.tgz</b></p> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p> <p>Path to dependency file: /functions/package.json</p> <p>Path to vulnerable library: /functions/node_modules/ansi-regex/package.json,/port/node_modules/ansi-regex/package.json,/port/functions/node_modules/ansi-regex/package.json,/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - eslint-7.17.0.tgz (Root Library) - strip-ansi-6.0.0.tgz - :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/husnuljahneer/PortfolioReact/commit/965058239ea8a371ffb53aab2e372f73b551779f">965058239ea8a371ffb53aab2e372f73b551779f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (eslint): 7.18.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"eslint","packageVersion":"7.17.0","packageFilePaths":["/functions/package.json"],"isTransitiveDependency":false,"dependencyTree":"eslint:7.17.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.18.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
eslint-7.17.0.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eslint-7.17.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /port/functions/package.json</p> <p>Path to vulnerable library: /functions/node_modules/ansi-regex/package.json,/port/node_modules/ansi-regex/package.json,/port/functions/node_modules/ansi-regex/package.json,/node_modules/ansi-regex/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/husnuljahneer/PortfolioReact/commit/965058239ea8a371ffb53aab2e372f73b551779f">965058239ea8a371ffb53aab2e372f73b551779f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-3807](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | ansi-regex-5.0.0.tgz | Transitive | 7.18.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3807</summary> ### Vulnerable Library - <b>ansi-regex-5.0.0.tgz</b></p> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p> <p>Path to dependency file: /functions/package.json</p> <p>Path to vulnerable library: /functions/node_modules/ansi-regex/package.json,/port/node_modules/ansi-regex/package.json,/port/functions/node_modules/ansi-regex/package.json,/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - eslint-7.17.0.tgz (Root Library) - strip-ansi-6.0.0.tgz - :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/husnuljahneer/PortfolioReact/commit/965058239ea8a371ffb53aab2e372f73b551779f">965058239ea8a371ffb53aab2e372f73b551779f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (eslint): 7.18.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"eslint","packageVersion":"7.17.0","packageFilePaths":["/functions/package.json"],"isTransitiveDependency":false,"dependencyTree":"eslint:7.17.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.18.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_process
eslint tgz vulnerabilities highest severity is vulnerable library eslint tgz path to dependency file port functions package json path to vulnerable library functions node modules ansi regex package json port node modules ansi regex package json port functions node modules ansi regex package json node modules ansi regex package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high ansi regex tgz transitive details cve vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file functions package json path to vulnerable library functions node modules ansi regex package json port node modules ansi regex package json port functions node modules ansi regex package json node modules ansi regex package json dependency hierarchy eslint tgz root library strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex direct dependency fix resolution eslint step up your open source security game with whitesource istransitivedependency false dependencytree eslint isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails ansi regex is vulnerable to inefficient regular expression complexity vulnerabilityurl
0
105,875
16,661,241,883
IssuesEvent
2021-06-06 11:08:46
AlexRogalskiy/weather-time
https://api.github.com/repos/AlexRogalskiy/weather-time
opened
CVE-2018-11698 (High) detected in node-sassv4.13.1
security vulnerability
## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.13.1</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/weather-time/commit/559d39620a54998a6f35e7cbea6c60528e5da3b0">559d39620a54998a6f35e7cbea6c60528e5da3b0</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>weather-time/node_modules/node-sass/src/libsass/src/sass_context.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: Libsass-3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11698 (High) detected in node-sassv4.13.1 - ## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.13.1</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/weather-time/commit/559d39620a54998a6f35e7cbea6c60528e5da3b0">559d39620a54998a6f35e7cbea6c60528e5da3b0</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>weather-time/node_modules/node-sass/src/libsass/src/sass_context.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: Libsass-3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in node cve high severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href vulnerable source files weather time node modules node sass src libsass src sass context cpp vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
36,692
15,043,459,388
IssuesEvent
2021-02-03 00:45:43
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
Incorrect squiggles after changing struct definition in header file
Language Service Visual Studio bug fixed (release pending)
Using the following code as the starting point, you can get IntelliSense to tell you that type T cannot be assigned to type T if you change the definition of type T in the header file. **1.h** ``` namespace N { class C { public: struct Inner { }; C(); private: Inner m_inner; }; }; ``` **1.cpp** ``` #include "1.h" using namespace N; C::C() { m_inner = Inner{}; } ``` The way I repro this is by opening both files in the editor, adding some fields to `struct Inner`, and then switching focus to 1.cpp. At that point, I get a squiggle on the equals sign telling me the assignment can't be done. It repros with AutoPCH disabled, it does not repro in VS nor if the class is defined outside of the namespace (as far as I can tell). It's not 100% repro, but it does appear quite frequently. Just keep adding/removing fields in the struct and some combination of saving/not saving the header file after the edits will repro the issue.
1.0
Incorrect squiggles after changing struct definition in header file - Using the following code as the starting point, you can get IntelliSense to tell you that type T cannot be assigned to type T if you change the definition of type T in the header file. **1.h** ``` namespace N { class C { public: struct Inner { }; C(); private: Inner m_inner; }; }; ``` **1.cpp** ``` #include "1.h" using namespace N; C::C() { m_inner = Inner{}; } ``` The way I repro this is by opening both files in the editor, adding some fields to `struct Inner`, and then switching focus to 1.cpp. At that point, I get a squiggle on the equals sign telling me the assignment can't be done. It repros with AutoPCH disabled, it does not repro in VS nor if the class is defined outside of the namespace (as far as I can tell). It's not 100% repro, but it does appear quite frequently. Just keep adding/removing fields in the struct and some combination of saving/not saving the header file after the edits will repro the issue.
non_process
incorrect squiggles after changing struct definition in header file using the following code as the starting point you can get intellisense to tell you that type t cannot be assigned to type t if you change the definition of type t in the header file h namespace n class c public struct inner c private inner m inner cpp include h using namespace n c c m inner inner the way i repro this is by opening both files in the editor adding some fields to struct inner and then switching focus to cpp at that point i get a squiggle on the equals sign telling me the assignment can t be done it repros with autopch disabled it does not repro in vs nor if the class is defined outside of the namespace as far as i can tell it s not repro but it does appear quite frequently just keep adding removing fields in the struct and some combination of saving not saving the header file after the edits will repro the issue
0
8,861
11,957,212,289
IssuesEvent
2020-04-04 13:36:30
martinlindhe/wmi_exporter
https://api.github.com/repos/martinlindhe/wmi_exporter
closed
How to get the Memory usage by process matching with Task manager
collector/process question
Tried `wmi_process_working_set`, `wmi_process_virtual_bytes` and `wmi_process_private_bytes` metrics under the [process collector](https://github.com/martinlindhe/wmi_exporter/blob/master/docs/collector.process.md), but none of them match with the result of Task manager. i.e. powershell mem usage from task manager is 32.1MB, but `wmi_process_working_set` -> 73285632 `wmi_process_virtual_bytes` -> 2204049985536 `wmi_process_private_bytes` -> 87359488
1.0
How to get the Memory usage by process matching with Task manager - Tried `wmi_process_working_set`, `wmi_process_virtual_bytes` and `wmi_process_private_bytes` metrics under the [process collector](https://github.com/martinlindhe/wmi_exporter/blob/master/docs/collector.process.md), but none of them match with the result of Task manager. i.e. powershell mem usage from task manager is 32.1MB, but `wmi_process_working_set` -> 73285632 `wmi_process_virtual_bytes` -> 2204049985536 `wmi_process_private_bytes` -> 87359488
process
how to get the memory usage by process matching with task manager tried wmi process working set wmi process virtual bytes and wmi process private bytes metrics under the but none of them match with the result of task manager i e powershell mem usage from task manager is but wmi process working set wmi process virtual bytes wmi process private bytes
1
16,340
20,999,904,501
IssuesEvent
2022-03-29 16:26:09
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
`binary-release` deploy step fails setting `NEXT_DEV_VERSION`
type: bug process: release stage: ready for work
This step consistently fails: https://github.com/cypress-io/cypress/blob/8c00988c2a867f8aa71d005ce99a7887210cbd98/DEPLOY.md#L162-L165 The download server manifest is updated correctly; however, there is an error when trying to update NEXT_DEV_VERSION in CI, so it has to be done manually for the time being.
1.0
`binary-release` deploy step fails setting `NEXT_DEV_VERSION` - This step consistently fails: https://github.com/cypress-io/cypress/blob/8c00988c2a867f8aa71d005ce99a7887210cbd98/DEPLOY.md#L162-L165 The download server manifest is updated correctly; however, there is an error when trying to update NEXT_DEV_VERSION in CI, so it has to be done manually for the time being.
process
binary release deploy step fails setting next dev version this step consistently fails the download server manifest is updated correctly however there is an error when trying to update next dev version in ci so it has to be done manually for the time being
1
7,427
10,546,152,500
IssuesEvent
2019-10-02 20:47:02
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Query expansion failed - Display custom foreign key value
Priority:P1 Query Processor Type:Bug
**Your browser and the version:** Firefox 63.0.3 (64 bits) **Your operating system:** Windows 10 **Your databases:** MySQL **Metabase version:** 0.30.4 **Metabase hosting environment:** Local ? Running the jar file from cmd. **Metabase internal database:** H2 **Steps to reproduce the issue:** 1. Create 3 tables (*Table 1*, *Table 2*, *Table 3*) 2. Create a foreign key between *Table 1* -> *Table 2* 3. Create a foreign key between *Table 2* -> *Table 3* 4. On Metabase, configure the foreign key of *Table 2* to display another value than the ID 5. Create a new question using the Table visualization, display any fields from *Table 1* and the field from *Table 2* which has the foreign key > Query expansion failed: could not find table 22 (FK ID = 175). Resolved tables ([fk-id table-id]): ([nil 23] [197 21]) **Step 4** is causing the problem, if I set back the displayed value to default (ID) I don't get the error.
1.0
Query expansion failed - Display custom foreign key value - **Your browser and the version:** Firefox 63.0.3 (64 bits) **Your operating system:** Windows 10 **Your databases:** MySQL **Metabase version:** 0.30.4 **Metabase hosting environment:** Local ? Running the jar file from cmd. **Metabase internal database:** H2 **Steps to reproduce the issue:** 1. Create 3 tables (*Table 1*, *Table 2*, *Table 3*) 2. Create a foreign key between *Table 1* -> *Table 2* 3. Create a foreign key between *Table 2* -> *Table 3* 4. On Metabase, configure the foreign key of *Table 2* to display another value than the ID 5. Create a new question using the Table visualization, display any fields from *Table 1* and the field from *Table 2* which has the foreign key > Query expansion failed: could not find table 22 (FK ID = 175). Resolved tables ([fk-id table-id]): ([nil 23] [197 21]) **Step 4** is causing the problem, if I set back the displayed value to default (ID) I don't get the error.
process
query expansion failed display custom foreign key value your browser and the version firefox bits your operating system windows your databases mysql metabase version metabase hosting environment local running the jar file from cmd metabase internal database steps to reproduce the issue create tables table table table create a foreign key between table table create a foreign key between table table on metabase configure the foreign key of table to display another value than the id create a new question using the table visualization display any fields from table and the field from table which has the foreign key query expansion failed could not find table fk id resolved tables step is causing the problem if i set back the displayed value to default id i don t get the error
1
18,050
24,058,273,687
IssuesEvent
2022-09-16 19:09:08
GoogleCloudPlatform/terraform-mean-cloudrun-mongodb
https://api.github.com/repos/GoogleCloudPlatform/terraform-mean-cloudrun-mongodb
reopened
Identify tasks to be automated
process
Referring to #3, identify the steps that need to be automated by Terraform.
1.0
Identify tasks to be automated - Referring to #3, identify the steps that need to be automated by Terraform.
process
identify tasks to be automated referring to identify the steps that need to be automated by terraform
1
300,007
22,636,414,420
IssuesEvent
2022-06-30 19:29:11
CUHealthAI/sqlite-clean
https://api.github.com/repos/CUHealthAI/sqlite-clean
opened
Provide short description and codes for referencing specific lint or fixes
documentation enhancement
Provide short descriptions (ex. `detect-this-thing`) and codes (ex. `L001` or `F001`) for each lint and fix involved with this repo to enable enhanced documentation and code-based references.
1.0
Provide short description and codes for referencing specific lint or fixes - Provide short descriptions (ex. `detect-this-thing`) and codes (ex. `L001` or `F001`) for each lint and fix involved with this repo to enable enhanced documentation and code-based references.
non_process
provide short description and codes for referencing specific lint or fixes provide short descriptions ex detect this thing and codes ex or for each lint and fix involved with this repo to enable enhanced documentation and code based references
0
5,007
7,840,784,795
IssuesEvent
2018-06-18 17:28:37
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
pubsub system test: "failed with exit code -11"
api: pubsub flaky testing type: process
Pubsub system tests run successfully, but [die occasionally](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6873) with a segfault(?) on exit: ``` nox > Running session system(py='2.7') nox > /usr/local/bin/python3.6 -m virtualenv /var/code/gcp/.nox/sys-2-7 -p python2.7 nox > chdir /var/code/gcp/pubsub nox > pip install --upgrade --pre grpcio nox > pip install --upgrade mock pytest ../api_core ../core nox > pip install --upgrade ../test_utils/ nox > pip install --upgrade . nox > py.test --quiet tests/system.py ... [100%] 3 passed in 30.45 seconds nox > Command py.test --quiet tests/system.py failed with exit code -11 nox > Session system(py='2.7') failed. ```
1.0
pubsub system test: "failed with exit code -11" - Pubsub system tests run successfully, but [die occasionally](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6873) with a segfault(?) on exit: ``` nox > Running session system(py='2.7') nox > /usr/local/bin/python3.6 -m virtualenv /var/code/gcp/.nox/sys-2-7 -p python2.7 nox > chdir /var/code/gcp/pubsub nox > pip install --upgrade --pre grpcio nox > pip install --upgrade mock pytest ../api_core ../core nox > pip install --upgrade ../test_utils/ nox > pip install --upgrade . nox > py.test --quiet tests/system.py ... [100%] 3 passed in 30.45 seconds nox > Command py.test --quiet tests/system.py failed with exit code -11 nox > Session system(py='2.7') failed. ```
process
pubsub system test failed with exit code pubsub system tests run successfully but with a segfault on exit nox running session system py nox usr local bin m virtualenv var code gcp nox sys p nox chdir var code gcp pubsub nox pip install upgrade pre grpcio nox pip install upgrade mock pytest api core core nox pip install upgrade test utils nox pip install upgrade nox py test quiet tests system py passed in seconds nox command py test quiet tests system py failed with exit code nox session system py failed
1
11,272
14,060,523,343
IssuesEvent
2020-11-03 06:13:20
slok/kahoy
https://api.github.com/repos/slok/kahoy
closed
Detect namespace/cluster scope resources using the discovery API
bug kubernetes processor resources
At this moment Kahoy isn't smart in that way. It loads resources in [`Unstructured`](https://godoc.org/github.com/kubernetes/apimachinery/pkg/apis/meta/v1/unstructured#Unstructured) manner (doesn't have the concept of specific resources and they identity). This has problems: - Don't know which resource type is cluster scoped or namespaced. - Resource IDs are based on resource type, name and namespace. - If a cluster scoped has the namespace defined and changes, to Kahoy this is a different resource, and shouldn't be. - This can trigger unexpected garbage collections. The solution, should be talking with the Kubernetes API and check the resource information to know which resources are namespaced and which not, this way setting `default` as namespace to all cluster scoped resources. Additional information: - https://godoc.org/k8s.io/client-go/discovery - https://godoc.org/k8s.io/apimachinery/pkg/apis/meta/v1#APIResource
1.0
Detect namespace/cluster scope resources using the discovery API - At this moment Kahoy isn't smart in that way. It loads resources in [`Unstructured`](https://godoc.org/github.com/kubernetes/apimachinery/pkg/apis/meta/v1/unstructured#Unstructured) manner (doesn't have the concept of specific resources and they identity). This has problems: - Don't know which resource type is cluster scoped or namespaced. - Resource IDs are based on resource type, name and namespace. - If a cluster scoped has the namespace defined and changes, to Kahoy this is a different resource, and shouldn't be. - This can trigger unexpected garbage collections. The solution, should be talking with the Kubernetes API and check the resource information to know which resources are namespaced and which not, this way setting `default` as namespace to all cluster scoped resources. Additional information: - https://godoc.org/k8s.io/client-go/discovery - https://godoc.org/k8s.io/apimachinery/pkg/apis/meta/v1#APIResource
process
detect namespace cluster scope resources using the discovery api at this moment kahoy isn t smart in that way it loads resources in manner doesn t have the concept of specific resources and they identity this has problems don t know which resource type is cluster scoped or namespaced resource ids are based on resource type name and namespace if a cluster scoped has the namespace defined and changes to kahoy this is a different resource and shouldn t be this can trigger unexpected garbage collections the solution should be talking with the kubernetes api and check the resource information to know which resources are namespaced and which not this way setting default as namespace to all cluster scoped resources additional information
1
26,166
6,755,440,057
IssuesEvent
2017-10-24 00:31:44
jascam/CodePlexFoo
https://api.github.com/repos/jascam/CodePlexFoo
closed
Create Example: CSCodeDOM
CodePlexMigrationInitiated enhancement impact: Medium
The CSCodeDOM project demonstrates how to use the .NET CodeDOM mechanism to enable dynamic souce code generation and compilation at runtime. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '3259' Vote count: '1'
1.0
Create Example: CSCodeDOM - The CSCodeDOM project demonstrates how to use the .NET CodeDOM mechanism to enable dynamic souce code generation and compilation at runtime. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '3259' Vote count: '1'
non_process
create example cscodedom the cscodedom project demonstrates how to use the net codedom mechanism to enable dynamic souce code generation and compilation at runtime migrated codeplex work item details codeplex work item id vote count
0
18,589
24,568,295,726
IssuesEvent
2022-10-13 06:22:27
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Failure message should be changed , which comes when participants are not eligible for the study in the standalone app
Bug P1 iOS Android Process: Fixed Process: Tested QA Process: Tested dev
**Steps:** 1. Install the standalone app 2. Sign up 3. Click on Participate button 4. Fail the eligibility process 5. Now, Verify the failure message **AR:** Getting message as attached in the below screenshot ![Android1](https://user-images.githubusercontent.com/86007179/165768472-a815c29b-21de-4593-82e7-94243ac9928e.png) **ER:** Failure message should be changed , which comes when participants are not eligible for the study in the standalone app
3.0
[Android] Failure message should be changed , which comes when participants are not eligible for the study in the standalone app - **Steps:** 1. Install the standalone app 2. Sign up 3. Click on Participate button 4. Fail the eligibility process 5. Now, Verify the failure message **AR:** Getting message as attached in the below screenshot ![Android1](https://user-images.githubusercontent.com/86007179/165768472-a815c29b-21de-4593-82e7-94243ac9928e.png) **ER:** Failure message should be changed , which comes when participants are not eligible for the study in the standalone app
process
failure message should be changed which comes when participants are not eligible for the study in the standalone app steps install the standalone app sign up click on participate button fail the eligibility process now verify the failure message ar getting message as attached in the below screenshot er failure message should be changed which comes when participants are not eligible for the study in the standalone app
1
20,566
27,225,410,453
IssuesEvent
2023-02-21 09:21:39
oasis-tcs/csaf
https://api.github.com/repos/oasis-tcs/csaf
closed
Register CSAF in security.txt
csaf 2.0 external oasis_tc_process
We should register CSAF as a dedicated field in the 'security.txt` ([Section 7.1.8](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#718-requirement-8-securitytxt); see also https://github.com/securitytxt/security-txt/issues/200). This ticket tracks the effort as it is the last action item open from #152.
1.0
Register CSAF in security.txt - We should register CSAF as a dedicated field in the 'security.txt` ([Section 7.1.8](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#718-requirement-8-securitytxt); see also https://github.com/securitytxt/security-txt/issues/200). This ticket tracks the effort as it is the last action item open from #152.
process
register csaf in security txt we should register csaf as a dedicated field in the security txt see also this ticket tracks the effort as it is the last action item open from
1
672,429
22,825,616,071
IssuesEvent
2022-07-12 08:19:12
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
rule34.xxx - desktop site instead of mobile site
browser-firefox-mobile nsfw priority-normal engine-gecko
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/107239 --> **URL**: https://rule34.xxx/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Safari **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: It Won't Let Me Pass. I Got To <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
rule34.xxx - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/107239 --> **URL**: https://rule34.xxx/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Safari **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: It Won't Let Me Pass. I Got To <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
xxx desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes safari problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce it won t let me pass i got to browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
70,209
15,057,340,915
IssuesEvent
2021-02-03 21:31:52
raft-tech/TANF-app
https://api.github.com/repos/raft-tech/TANF-app
closed
As a user, I can view a placeholder page that is only tied to my role
DUPE dev frontend security
Description: This issue will deliver OFA admin or test users acting as STT being able to view a placeholder page that is only associated with their role. AC: - [ ] Authenticated user views a placeholder page with their email address and assigned role Frontend - [ ] Add user role text with dynamic role (e.g. "Hi, you're a data prepper!" or Hi, you're an OFA Admin!") - [ ] Add card with role information - [ ] Add tests to cover Open Questions: - What roles are going to be populated based for OFA MVP? - STT user (Test user) and OFA Admiin - What security controls does this impact? @carltonsmith to add Note: - One page that will be dynamically served based on the role. - Signed out button is not needed since user can sign out top right hand side of the page. Supporting Documentation: ![Screen Shot 2020-10-20 at 8.38.23 PM.png](https://images.zenhubusercontent.com/5f18d5aef8f52808880b2109/6b25b4fa-8a0d-43fe-87e0-f1915231fe33)
True
As a user, I can view a placeholder page that is only tied to my role - Description: This issue will deliver OFA admin or test users acting as STT being able to view a placeholder page that is only associated with their role. AC: - [ ] Authenticated user views a placeholder page with their email address and assigned role Frontend - [ ] Add user role text with dynamic role (e.g. "Hi, you're a data prepper!" or Hi, you're an OFA Admin!") - [ ] Add card with role information - [ ] Add tests to cover Open Questions: - What roles are going to be populated based for OFA MVP? - STT user (Test user) and OFA Admiin - What security controls does this impact? @carltonsmith to add Note: - One page that will be dynamically served based on the role. - Signed out button is not needed since user can sign out top right hand side of the page. Supporting Documentation: ![Screen Shot 2020-10-20 at 8.38.23 PM.png](https://images.zenhubusercontent.com/5f18d5aef8f52808880b2109/6b25b4fa-8a0d-43fe-87e0-f1915231fe33)
non_process
as a user i can view a placeholder page that is only tied to my role description this issue will deliver ofa admin or test users acting as stt being able to view a placeholder page that is only associated with their role ac authenticated user views a placeholder page with their email address and assigned role frontend add user role text with dynamic role e g hi you re a data prepper or hi you re an ofa admin add card with role information add tests to cover open questions what roles are going to be populated based for ofa mvp stt user test user and ofa admiin what security controls does this impact carltonsmith to add note one page that will be dynamically served based on the role signed out button is not needed since user can sign out top right hand side of the page supporting documentation
0
4,392
7,285,471,377
IssuesEvent
2018-02-23 04:24:58
BlesseNtumble/GalaxySpace
https://api.github.com/repos/BlesseNtumble/GalaxySpace
closed
Crash on changing dimensions in a rocket
in the process of correcting is a known bug
GalactiCraft 502 Galaxy Space 1.1.8 Forge 10.13.4.1614 ---- Minecraft Crash Report ---- // On the bright side, I bought you a teddy bear! Time: 27/09/17 16:27 Description: Exception generating new chunk java.lang.NullPointerException: Exception generating new chunk at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.replaceBlocksForBiome(Unknown Source) at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.func_73154_d(Unknown Source) at net.minecraft.world.gen.ChunkProviderServer.originalLoadChunk(ChunkProviderServer.java:172) at net.minecraft.world.gen.ChunkProviderServer.loadChunk(ChunkProviderServer.java:131) at net.minecraft.world.gen.ChunkProviderServer.func_73158_c(ChunkProviderServer.java:101) at net.minecraft.world.gen.ChunkProviderServer.func_73154_d(ChunkProviderServer.java:199) at net.minecraft.world.World.func_72964_e(World.java:419) at net.minecraft.world.SpawnerAnimals.func_77192_a(SpawnerAnimals.java:144) at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:161) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:625) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:547) at net.minecraft.server.integrated.IntegratedServer.func_71217_p(IntegratedServer.java:186) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:427) at net.minecraft.server.MinecraftServer$2.run(MinecraftServer.java:685) A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Stacktrace: at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.replaceBlocksForBiome(Unknown Source) at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.func_73154_d(Unknown Source) -- Chunk to be generated -- Details: Location: 6,10 Position hash: 42949672966 Generator: RandomLevelSource Stacktrace: at net.minecraft.world.gen.ChunkProviderServer.originalLoadChunk(ChunkProviderServer.java:172) at net.minecraft.world.gen.ChunkProviderServer.loadChunk(ChunkProviderServer.java:131) at net.minecraft.world.gen.ChunkProviderServer.func_73158_c(ChunkProviderServer.java:101) at net.minecraft.world.gen.ChunkProviderServer.func_73154_d(ChunkProviderServer.java:199) at net.minecraft.world.World.func_72964_e(World.java:419) at net.minecraft.world.SpawnerAnimals.func_77192_a(SpawnerAnimals.java:144) at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:161) -- Affected level -- Details: Level name: moontest All players: 1 total; [GCEntityPlayerMP['TiranDirth'/25930, l='moontest', x=193.50, y=900.00, z=155.50]] Chunk stats: ServerChunkCache: 1 Drop: 0 Level seed: -1001766155814337237 Level generator: ID 00 - default, ver 1. Features enabled: false Level generator options: Level spawn location: World: (0,0,0), Chunk: (at 0,0,0 in 0,0; contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511) Level time: 0 game time, 0 day time Level dimension: 0 Level storage version: 0x00000 - Unknown? Level weather: Rain time: 0 (now: false), thunder time: 0 (now: false) Level game mode: ~~ERROR~~ NullPointerException: null Stacktrace: at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:625) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:547) at net.minecraft.server.integrated.IntegratedServer.func_71217_p(IntegratedServer.java:186) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:427) at net.minecraft.server.MinecraftServer$2.run(MinecraftServer.java:685) -- System Details -- Details: Minecraft Version: 1.7.10 Operating System: Windows 10 (amd64) version 10.0 Java Version: 1.8.0_144, Oracle Corporation Java VM Version: Java HotSpot(TM) 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 4221403136 bytes (4025 MB) / 6442450944 bytes (6144 MB) up to 6442450944 bytes (6144 MB) JVM Flags: 5 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xms6144m -Xmx6144m -XX:+UseG1GC -XX:MaxGCPauseMillis=4 AABB Pool Size: 0 (0 bytes; 0 MB) allocated, 0 (0 bytes; 0 MB) used IntCache: cache: 10, tcache: 53, allocated: 3, tallocated: 42 FML: MCP v9.05 FML v7.10.99.99 Minecraft Forge 10.13.4.1614 Optifine OptiFine_1.7.10_HD_U_D8 76 mods loaded, 76 mods active States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored UCHIJAAAAAAAAA mcp{9.05} [Minecraft Coder Pack] (minecraft.jar) UCHIJAAAAAAAAA FML{7.10.99.99} [Forge Mod Loader] (modpack.jar) UCHIJAAAAAAAAA Forge{10.13.4.1614} [Minecraft Forge] (modpack.jar) UCHIJAAAAAAAAA appliedenergistics2-core{rv3-beta-6} [Applied Energistics 2 Core] (minecraft.jar) UCHIJAAAAAAAAA CodeChickenCore{1.0.7.47} [CodeChicken Core] (minecraft.jar) UCHIJAAAAAAAAA Micdoodlecore{} [Micdoodle8 Core] (minecraft.jar) UCHIJAAAAAAAAA NotEnoughItems{1.0.5.120} [Not Enough Items] (NotEnoughItems-1.7.10-1.0.5.120-universal.jar) UCHIJAAAAAAAAA OpenModsCore{0.10} [OpenModsCore] (minecraft.jar) UCHIJAAAAAAAAA <CoFH ASM>{000} [CoFH ASM] (minecraft.jar) UCHIJAAAAAAAAA <DragonAPI ASM>{0} [DragonAPI ASM Data Initialization] (minecraft.jar) UCHIJAAAAAAAAA appliedenergistics2{rv3-beta-6} [Applied Energistics 2] (appliedenergistics2-rv3-beta-6.jar) UCHIJAAAAAAAAA Baubles{1.0.1.10} [Baubles] (Baubles-1.7.10-1.0.1.10.jar) UCHIJAAAAAAAAA BiblioCraft{1.11.7} [BiblioCraft] (BiblioCraft[v1.11.7][MC1.7.10].jar) UCHIJAAAAAAAAA Botania{r1.8-249} [Botania] (Botania r1.8-249.jar) UCHIJAAAAAAAAA CoFHCore{1.7.10R3.1.4} [CoFH Core] (CoFHCore-[1.7.10]3.1.4-329.jar) UCHIJAAAAAAAAA BuildCraft|Core{7.1.22} [BuildCraft] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Transport{7.1.22} [BC Transport] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Factory{7.1.22} [BC Factory] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Silicon{7.1.22} [BC Silicon] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Robotics{7.1.22} [BC Robotics] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Energy{7.1.22} [BC Energy] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Builders{7.1.22} [BC Builders] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA Railcraft{9.12.2.0} [Railcraft] (Railcraft_1.7.10-9.12.2.0.jar) UCHIJAAAAAAAAA TwilightForest{2.3.7} [The Twilight Forest] (twilightforest-1.7.10-2.3.7.jar) UCHIJAAAAAAAAA ForgeMultipart{1.2.0.347} [Forge Multipart] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA chisel{2.9.5.11} [Chisel] (Chisel-2.9.5.11.jar) UCHIJAAAAAAAAA CarpentersBlocks{3.3.8.1} [Carpenter's Blocks] (Carpenter's Blocks v3.3.8.1 - MC 1.7.10.jar) UCHIJAAAAAAAAA HardcoreEnderExpansion{1.8.6} [Hardcore Ender Expansion] (HardcoreEnderExpansion MC-1.7.10 v1.8.6.jar) UCHIJAAAAAAAAA Mantle{1.7.10-0.3.2.jenkins191} [Mantle] (Mantle-1.7.10-0.3.2b.jar) UCHIJAAAAAAAAA Natura{2.2.0} [Natura] (natura-1.7.10-2.2.1a2.jar) UCHIJAAAAAAAAA Forestry{4.2.16.64} [Forestry for Minecraft] (forestry_1.7.10-4.2.16.64.jar) UCHIJAAAAAAAAA MrTJPCoreMod{1.1.0.33} [MrTJPCore] (MrTJPCore-1.7.10-1.1.0.33-universal.jar) UCHIJAAAAAAAAA ProjRed|Core{4.7.0pre12.95} [ProjectRed Core] (ProjectRed-1.7.10-4.7.0pre12.95-Base.jar) UCHIJAAAAAAAAA ThermalFoundation{1.7.10R1.2.6} [Thermal Foundation] (ThermalFoundation-[1.7.10]1.2.6-118.jar) UCHIJAAAAAAAAA ExtraUtilities{1.2.12} [Extra Utilities] (extrautilities-1.2.12.jar) UCHIJAAAAAAAAA ThermalExpansion{1.7.10R4.1.5} [Thermal Expansion] (ThermalExpansion-[1.7.10]4.1.5-248.jar) UCHIJAAAAAAAAA Waila{1.5.10} [Waila] (Waila-1.5.10_1.7.10.jar) UCHIJAAAAAAAAA TConstruct{1.7.10-1.8.8.build988} [Tinkers' Construct] (TConstruct-1.7.10-1.8.8.jar) UCHIJAAAAAAAAA DragonAPI{v18b} [DragonAPI] (DragonAPI 1.7.10 V18b.jar) UCHIJAAAAAAAAA ChromatiCraft{v18b} [ChromatiCraft] (ChromatiCraft 1.7.10 V18b.jar) UCHIJAAAAAAAAA DamageIndicatorsMod{3.2.0} [Damage Indicators] (Damage-Indicators-Mod-1.7.10.jar) UCHIJAAAAAAAAA endercore{1.7.10-0.2.0.39_beta} [EnderCore] (EnderCore-1.7.10-0.2.0.39_beta.jar) UCHIJAAAAAAAAA EnderIO{1.7.10-2.3.0.429_beta} [Ender IO] (EnderIO-1.7.10-2.3.0.429_beta.jar) UCHIJAAAAAAAAA extracells{2.3.14} [Extra Cells 2] (ExtraCells-1.7.10-2.3.14b197.jar) UCHIJAAAAAAAAA MetallurgyCore{4.0.4} [Metallurgy Core] (MetallurgyCore-1.7.10-4.0.4.18.jar) UCHIJAAAAAAAAA Metallurgy{4.0.6} [Metallurgy 4] (Metallurgy-1.7.10-4.0.6.80.jar) UCHIJAAAAAAAAA ExtraTiC{1.4.6} [ExtraTiC] (ExtraTiC-1.7.10-1.4.6.jar) UCHIJAAAAAAAAA GalacticraftCore{3.0.12} [Galacticraft Core] (GalacticraftCore-1.7-3.0.12.502.jar) UCHIJAAAAAAAAA GalacticraftMars{3.0.12} [Galacticraft Planets] (Galacticraft-Planets-1.7-3.0.12.502.jar) UCHIJAAAAAAAAA GalaxySpace{1.1.8} [GalaxySpace] (GalaxySpace-1.1.8 STABLE.jar) UCHIJAAAAAAAAA GeoStrata{v18a} [GeoStrata] (GeoStrata 1.7.10 V18a.jar) UCHIJAAAAAAAAA iChunUtil{4.2.3} [iChunUtil] (iChunUtil-4.2.3.jar) UCHIJAAAAAAAAA Hats{4.0.1} [Hats] (Hats-4.0.1.jar) UCHIJAAAAAAAAA inpure|core{1.7.10R1.0.0B9} [INpureCore] (INpureCore-[1.7.10]1.0.0B9-62.jar) UCHIJAAAAAAAAA inventorytweaks{1.59-dev-152-cf6e263} [Inventory Tweaks] (InventoryTweaks-1.59-dev-152.jar) UCHIJAAAAAAAAA journeymap{5.1.4p2} [JourneyMap] (journeymap-1.7.10-5.1.4p2-unlimited.jar) UCHIJAAAAAAAAA Mekanism{9.1.0} [Mekanism] (Mekanism-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MekanismGenerators{9.1.0} [MekanismGenerators] (MekanismGenerators-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MekanismTools{9.1.0} [MekanismTools] (MekanismTools-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MineTweaker3{3.0.10} [MineTweaker 3] (MineTweaker3-1.7.10-3.0.10B.jar) UCHIJAAAAAAAAA MTRM{1.0} [MineTweakerRecipeMaker] (MineTweakerRecipeMaker-1.7.10-1.1.0.11.jar) UCHIJAAAAAAAAA modtweaker2{0.9.6} [Mod Tweaker 2] (ModTweaker2-0.9.6.jar) UCHIJAAAAAAAAA Morpheus{1.7.10-1.5.26} [Morpheus] (Morpheus-1.7.10-1.5.26.jar) UCHIJAAAAAAAAA OpenMods{0.10} [OpenMods] (OpenModsLib-1.7.10-0.10.jar) UCHIJAAAAAAAAA OpenBlocks{1.6} [OpenBlocks] (OpenBlocks-1.7.10-1.6.jar) UCHIJAAAAAAAAA ProjRed|Integration{4.7.0pre12.95} [ProjectRed Integration] (ProjectRed-1.7.10-4.7.0pre12.95-Integration.jar) UCHIJAAAAAAAAA ProjRed|Transmission{4.7.0pre12.95} [ProjectRed Transmission] (ProjectRed-1.7.10-4.7.0pre12.95-Integration.jar) UCHIJAAAAAAAAA ProjRed|Illumination{4.7.0pre12.95} [ProjectRed Illumination] (ProjectRed-1.7.10-4.7.0pre12.95-Lighting.jar) UCHIJAAAAAAAAA ProjRed|Exploration{4.7.0pre12.95} [ProjectRed Exploration] (ProjectRed-1.7.10-4.7.0pre12.95-World.jar) UCHIJAAAAAAAAA ThermalDynamics{1.7.10R1.2.1} [Thermal Dynamics] (ThermalDynamics-[1.7.10]1.2.1-172.jar) UCHIJAAAAAAAAA WailaHarvestability{1.1.6} [Waila Harvestability] (WailaHarvestability-mc1.7.10-1.1.6.jar) UCHIJAAAAAAAAA witchery{0.24.1} [Witchery] (witchery-1.7.10-0.24.1.jar) UCHIJAAAAAAAAA ForgeMicroblock{1.2.0.347} [Forge Microblocks] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA McMultipart{1.2.0.347} [Minecraft Multipart Plugin] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA aobd{2.9.2} [Another One Bites The Dust] (AOBD-2.9.2.jar) UCHIJAAAAAAAAA IguanaTweaksTConstruct{1.7.10-2.1.6.163} [Iguana Tinker Tweaks] (IguanaTinkerTweaks-1.7.10-2.1.6.jar) GL info: ~~ERROR~~ RuntimeException: No OpenGL context found in the current thread. OpenModsLib class transformers: [stencil_patches:FINISHED],[movement_callback:FINISHED],[player_damage_hook:FINISHED],[map_gen_fix:FINISHED],[gl_capabilities_hook:FINISHED],[player_render_hook:FINISHED] Class transformer null safety: all safe AE2 Version: beta rv3-beta-6 for Forge 10.13.4.1448 CoFHCore: -[1.7.10]3.1.4-329 Mantle Environment: DO NOT REPORT THIS CRASH! Unsupported mods in environment: optifine ThermalFoundation: -[1.7.10]1.2.6-118 ThermalExpansion: -[1.7.10]4.1.5-248 TConstruct Environment: Environment healthy. ThermalDynamics: -[1.7.10]1.2.1-172 Chisel: Errors like "[FML]: Unable to lookup ..." are NOT the cause of this crash. You can safely ignore these errors. And update forge while you're at it. EnderIO: Found the following problem(s) with your installation: * Optifine is installed. This is NOT supported. * An unknown AE2 API is installed (rv3 from appliedenergistics2-rv3-beta-6.jar). Ender IO was build against API version rv2 and may or may not work with a newer version. * The RF API that is being used (1.7.10R1.3.1 from <unknown>) differes from that that is reported as being loaded (1.7.10R1.0.2 from DragonAPI 1.7.10 V18b.jar). It is a supported version, but that difference may lead to problems. This may have caused the error. Try reproducing the crash WITHOUT this/these mod(s) before reporting it. Stencil buffer state: Function set: GL30, pool: forge, bits: 8 Forestry : Warning: You have mods that change the behavior of Minecraft, ForgeModLoader, and/or Minecraft Forge to your client: Optifine These may have caused this error, and may not be supported. Try reproducing the crash WITHOUT these mods, and report it then. AE2 Integration: IC2:OFF, RotaryCraft:OFF, RC:ON, BuildCraftCore:ON, BuildCraftTransport:ON, BuildCraftBuilder:ON, RF:ON, RFItem:ON, MFR:OFF, DSU:ON, FZ:OFF, FMP:ON, RB:OFF, CLApi:OFF, Waila:ON, InvTweaks:ON, NEI:ON, CraftGuide:OFF, Mekanism:ON, ImmibisMicroblocks:OFF, BetterStorage:OFF, OpenComputers:OFF, PneumaticCraft:OFF Profiler Position: N/A (disabled) Vec3 Pool Size: 0 (0 bytes; 0 MB) allocated, 0 (0 bytes; 0 MB) used Player Count: 1 / 8; [GCEntityPlayerMP['NAMEREMOVED'/25930, l='moontest', x=193.50, y=900.00, z=155.50]] Type: Integrated Server (map_client.txt) Is Modded: Definitely; Client brand changed to 'fml,forge' OptiFine Version: OptiFine_1.7.10_HD_U_D8 Render Distance Chunks: 16 Mipmaps: 4 Anisotropic Filtering: 1 Antialiasing: 0 Multitexture: false Shaders: null OpenGlVersion: 4.5.13474 Compatibility Profile Context 22.19.162.4 OpenGlRenderer: AMD Radeon R9 200 Series OpenGlVendor: ATI Technologies Inc. CpuCount: 4 Happy to provide any more info if I can.
1.0
Crash on changing dimensions in a rocket - GalactiCraft 502 Galaxy Space 1.1.8 Forge 10.13.4.1614 ---- Minecraft Crash Report ---- // On the bright side, I bought you a teddy bear! Time: 27/09/17 16:27 Description: Exception generating new chunk java.lang.NullPointerException: Exception generating new chunk at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.replaceBlocksForBiome(Unknown Source) at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.func_73154_d(Unknown Source) at net.minecraft.world.gen.ChunkProviderServer.originalLoadChunk(ChunkProviderServer.java:172) at net.minecraft.world.gen.ChunkProviderServer.loadChunk(ChunkProviderServer.java:131) at net.minecraft.world.gen.ChunkProviderServer.func_73158_c(ChunkProviderServer.java:101) at net.minecraft.world.gen.ChunkProviderServer.func_73154_d(ChunkProviderServer.java:199) at net.minecraft.world.World.func_72964_e(World.java:419) at net.minecraft.world.SpawnerAnimals.func_77192_a(SpawnerAnimals.java:144) at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:161) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:625) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:547) at net.minecraft.server.integrated.IntegratedServer.func_71217_p(IntegratedServer.java:186) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:427) at net.minecraft.server.MinecraftServer$2.run(MinecraftServer.java:685) A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Stacktrace: at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.replaceBlocksForBiome(Unknown Source) at galaxyspace.SolarSystem.core.world.gen.ChunkProviderSpaceCraters.func_73154_d(Unknown Source) -- Chunk to be generated -- Details: Location: 6,10 Position hash: 42949672966 Generator: RandomLevelSource Stacktrace: at net.minecraft.world.gen.ChunkProviderServer.originalLoadChunk(ChunkProviderServer.java:172) at net.minecraft.world.gen.ChunkProviderServer.loadChunk(ChunkProviderServer.java:131) at net.minecraft.world.gen.ChunkProviderServer.func_73158_c(ChunkProviderServer.java:101) at net.minecraft.world.gen.ChunkProviderServer.func_73154_d(ChunkProviderServer.java:199) at net.minecraft.world.World.func_72964_e(World.java:419) at net.minecraft.world.SpawnerAnimals.func_77192_a(SpawnerAnimals.java:144) at net.minecraft.world.WorldServer.func_72835_b(WorldServer.java:161) -- Affected level -- Details: Level name: moontest All players: 1 total; [GCEntityPlayerMP['TiranDirth'/25930, l='moontest', x=193.50, y=900.00, z=155.50]] Chunk stats: ServerChunkCache: 1 Drop: 0 Level seed: -1001766155814337237 Level generator: ID 00 - default, ver 1. Features enabled: false Level generator options: Level spawn location: World: (0,0,0), Chunk: (at 0,0,0 in 0,0; contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511) Level time: 0 game time, 0 day time Level dimension: 0 Level storage version: 0x00000 - Unknown? Level weather: Rain time: 0 (now: false), thunder time: 0 (now: false) Level game mode: ~~ERROR~~ NullPointerException: null Stacktrace: at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:625) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:547) at net.minecraft.server.integrated.IntegratedServer.func_71217_p(IntegratedServer.java:186) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:427) at net.minecraft.server.MinecraftServer$2.run(MinecraftServer.java:685) -- System Details -- Details: Minecraft Version: 1.7.10 Operating System: Windows 10 (amd64) version 10.0 Java Version: 1.8.0_144, Oracle Corporation Java VM Version: Java HotSpot(TM) 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 4221403136 bytes (4025 MB) / 6442450944 bytes (6144 MB) up to 6442450944 bytes (6144 MB) JVM Flags: 5 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xms6144m -Xmx6144m -XX:+UseG1GC -XX:MaxGCPauseMillis=4 AABB Pool Size: 0 (0 bytes; 0 MB) allocated, 0 (0 bytes; 0 MB) used IntCache: cache: 10, tcache: 53, allocated: 3, tallocated: 42 FML: MCP v9.05 FML v7.10.99.99 Minecraft Forge 10.13.4.1614 Optifine OptiFine_1.7.10_HD_U_D8 76 mods loaded, 76 mods active States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored UCHIJAAAAAAAAA mcp{9.05} [Minecraft Coder Pack] (minecraft.jar) UCHIJAAAAAAAAA FML{7.10.99.99} [Forge Mod Loader] (modpack.jar) UCHIJAAAAAAAAA Forge{10.13.4.1614} [Minecraft Forge] (modpack.jar) UCHIJAAAAAAAAA appliedenergistics2-core{rv3-beta-6} [Applied Energistics 2 Core] (minecraft.jar) UCHIJAAAAAAAAA CodeChickenCore{1.0.7.47} [CodeChicken Core] (minecraft.jar) UCHIJAAAAAAAAA Micdoodlecore{} [Micdoodle8 Core] (minecraft.jar) UCHIJAAAAAAAAA NotEnoughItems{1.0.5.120} [Not Enough Items] (NotEnoughItems-1.7.10-1.0.5.120-universal.jar) UCHIJAAAAAAAAA OpenModsCore{0.10} [OpenModsCore] (minecraft.jar) UCHIJAAAAAAAAA <CoFH ASM>{000} [CoFH ASM] (minecraft.jar) UCHIJAAAAAAAAA <DragonAPI ASM>{0} [DragonAPI ASM Data Initialization] (minecraft.jar) UCHIJAAAAAAAAA appliedenergistics2{rv3-beta-6} [Applied Energistics 2] (appliedenergistics2-rv3-beta-6.jar) UCHIJAAAAAAAAA Baubles{1.0.1.10} [Baubles] (Baubles-1.7.10-1.0.1.10.jar) UCHIJAAAAAAAAA BiblioCraft{1.11.7} [BiblioCraft] (BiblioCraft[v1.11.7][MC1.7.10].jar) UCHIJAAAAAAAAA Botania{r1.8-249} [Botania] (Botania r1.8-249.jar) UCHIJAAAAAAAAA CoFHCore{1.7.10R3.1.4} [CoFH Core] (CoFHCore-[1.7.10]3.1.4-329.jar) UCHIJAAAAAAAAA BuildCraft|Core{7.1.22} [BuildCraft] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Transport{7.1.22} [BC Transport] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Factory{7.1.22} [BC Factory] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Silicon{7.1.22} [BC Silicon] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Robotics{7.1.22} [BC Robotics] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Energy{7.1.22} [BC Energy] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA BuildCraft|Builders{7.1.22} [BC Builders] (buildcraft-7.1.22.jar) UCHIJAAAAAAAAA Railcraft{9.12.2.0} [Railcraft] (Railcraft_1.7.10-9.12.2.0.jar) UCHIJAAAAAAAAA TwilightForest{2.3.7} [The Twilight Forest] (twilightforest-1.7.10-2.3.7.jar) UCHIJAAAAAAAAA ForgeMultipart{1.2.0.347} [Forge Multipart] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA chisel{2.9.5.11} [Chisel] (Chisel-2.9.5.11.jar) UCHIJAAAAAAAAA CarpentersBlocks{3.3.8.1} [Carpenter's Blocks] (Carpenter's Blocks v3.3.8.1 - MC 1.7.10.jar) UCHIJAAAAAAAAA HardcoreEnderExpansion{1.8.6} [Hardcore Ender Expansion] (HardcoreEnderExpansion MC-1.7.10 v1.8.6.jar) UCHIJAAAAAAAAA Mantle{1.7.10-0.3.2.jenkins191} [Mantle] (Mantle-1.7.10-0.3.2b.jar) UCHIJAAAAAAAAA Natura{2.2.0} [Natura] (natura-1.7.10-2.2.1a2.jar) UCHIJAAAAAAAAA Forestry{4.2.16.64} [Forestry for Minecraft] (forestry_1.7.10-4.2.16.64.jar) UCHIJAAAAAAAAA MrTJPCoreMod{1.1.0.33} [MrTJPCore] (MrTJPCore-1.7.10-1.1.0.33-universal.jar) UCHIJAAAAAAAAA ProjRed|Core{4.7.0pre12.95} [ProjectRed Core] (ProjectRed-1.7.10-4.7.0pre12.95-Base.jar) UCHIJAAAAAAAAA ThermalFoundation{1.7.10R1.2.6} [Thermal Foundation] (ThermalFoundation-[1.7.10]1.2.6-118.jar) UCHIJAAAAAAAAA ExtraUtilities{1.2.12} [Extra Utilities] (extrautilities-1.2.12.jar) UCHIJAAAAAAAAA ThermalExpansion{1.7.10R4.1.5} [Thermal Expansion] (ThermalExpansion-[1.7.10]4.1.5-248.jar) UCHIJAAAAAAAAA Waila{1.5.10} [Waila] (Waila-1.5.10_1.7.10.jar) UCHIJAAAAAAAAA TConstruct{1.7.10-1.8.8.build988} [Tinkers' Construct] (TConstruct-1.7.10-1.8.8.jar) UCHIJAAAAAAAAA DragonAPI{v18b} [DragonAPI] (DragonAPI 1.7.10 V18b.jar) UCHIJAAAAAAAAA ChromatiCraft{v18b} [ChromatiCraft] (ChromatiCraft 1.7.10 V18b.jar) UCHIJAAAAAAAAA DamageIndicatorsMod{3.2.0} [Damage Indicators] (Damage-Indicators-Mod-1.7.10.jar) UCHIJAAAAAAAAA endercore{1.7.10-0.2.0.39_beta} [EnderCore] (EnderCore-1.7.10-0.2.0.39_beta.jar) UCHIJAAAAAAAAA EnderIO{1.7.10-2.3.0.429_beta} [Ender IO] (EnderIO-1.7.10-2.3.0.429_beta.jar) UCHIJAAAAAAAAA extracells{2.3.14} [Extra Cells 2] (ExtraCells-1.7.10-2.3.14b197.jar) UCHIJAAAAAAAAA MetallurgyCore{4.0.4} [Metallurgy Core] (MetallurgyCore-1.7.10-4.0.4.18.jar) UCHIJAAAAAAAAA Metallurgy{4.0.6} [Metallurgy 4] (Metallurgy-1.7.10-4.0.6.80.jar) UCHIJAAAAAAAAA ExtraTiC{1.4.6} [ExtraTiC] (ExtraTiC-1.7.10-1.4.6.jar) UCHIJAAAAAAAAA GalacticraftCore{3.0.12} [Galacticraft Core] (GalacticraftCore-1.7-3.0.12.502.jar) UCHIJAAAAAAAAA GalacticraftMars{3.0.12} [Galacticraft Planets] (Galacticraft-Planets-1.7-3.0.12.502.jar) UCHIJAAAAAAAAA GalaxySpace{1.1.8} [GalaxySpace] (GalaxySpace-1.1.8 STABLE.jar) UCHIJAAAAAAAAA GeoStrata{v18a} [GeoStrata] (GeoStrata 1.7.10 V18a.jar) UCHIJAAAAAAAAA iChunUtil{4.2.3} [iChunUtil] (iChunUtil-4.2.3.jar) UCHIJAAAAAAAAA Hats{4.0.1} [Hats] (Hats-4.0.1.jar) UCHIJAAAAAAAAA inpure|core{1.7.10R1.0.0B9} [INpureCore] (INpureCore-[1.7.10]1.0.0B9-62.jar) UCHIJAAAAAAAAA inventorytweaks{1.59-dev-152-cf6e263} [Inventory Tweaks] (InventoryTweaks-1.59-dev-152.jar) UCHIJAAAAAAAAA journeymap{5.1.4p2} [JourneyMap] (journeymap-1.7.10-5.1.4p2-unlimited.jar) UCHIJAAAAAAAAA Mekanism{9.1.0} [Mekanism] (Mekanism-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MekanismGenerators{9.1.0} [MekanismGenerators] (MekanismGenerators-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MekanismTools{9.1.0} [MekanismTools] (MekanismTools-1.7.10-9.1.0.281.jar) UCHIJAAAAAAAAA MineTweaker3{3.0.10} [MineTweaker 3] (MineTweaker3-1.7.10-3.0.10B.jar) UCHIJAAAAAAAAA MTRM{1.0} [MineTweakerRecipeMaker] (MineTweakerRecipeMaker-1.7.10-1.1.0.11.jar) UCHIJAAAAAAAAA modtweaker2{0.9.6} [Mod Tweaker 2] (ModTweaker2-0.9.6.jar) UCHIJAAAAAAAAA Morpheus{1.7.10-1.5.26} [Morpheus] (Morpheus-1.7.10-1.5.26.jar) UCHIJAAAAAAAAA OpenMods{0.10} [OpenMods] (OpenModsLib-1.7.10-0.10.jar) UCHIJAAAAAAAAA OpenBlocks{1.6} [OpenBlocks] (OpenBlocks-1.7.10-1.6.jar) UCHIJAAAAAAAAA ProjRed|Integration{4.7.0pre12.95} [ProjectRed Integration] (ProjectRed-1.7.10-4.7.0pre12.95-Integration.jar) UCHIJAAAAAAAAA ProjRed|Transmission{4.7.0pre12.95} [ProjectRed Transmission] (ProjectRed-1.7.10-4.7.0pre12.95-Integration.jar) UCHIJAAAAAAAAA ProjRed|Illumination{4.7.0pre12.95} [ProjectRed Illumination] (ProjectRed-1.7.10-4.7.0pre12.95-Lighting.jar) UCHIJAAAAAAAAA ProjRed|Exploration{4.7.0pre12.95} [ProjectRed Exploration] (ProjectRed-1.7.10-4.7.0pre12.95-World.jar) UCHIJAAAAAAAAA ThermalDynamics{1.7.10R1.2.1} [Thermal Dynamics] (ThermalDynamics-[1.7.10]1.2.1-172.jar) UCHIJAAAAAAAAA WailaHarvestability{1.1.6} [Waila Harvestability] (WailaHarvestability-mc1.7.10-1.1.6.jar) UCHIJAAAAAAAAA witchery{0.24.1} [Witchery] (witchery-1.7.10-0.24.1.jar) UCHIJAAAAAAAAA ForgeMicroblock{1.2.0.347} [Forge Microblocks] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA McMultipart{1.2.0.347} [Minecraft Multipart Plugin] (ForgeMultipart-1.7.10-1.2.0.347-universal.jar) UCHIJAAAAAAAAA aobd{2.9.2} [Another One Bites The Dust] (AOBD-2.9.2.jar) UCHIJAAAAAAAAA IguanaTweaksTConstruct{1.7.10-2.1.6.163} [Iguana Tinker Tweaks] (IguanaTinkerTweaks-1.7.10-2.1.6.jar) GL info: ~~ERROR~~ RuntimeException: No OpenGL context found in the current thread. OpenModsLib class transformers: [stencil_patches:FINISHED],[movement_callback:FINISHED],[player_damage_hook:FINISHED],[map_gen_fix:FINISHED],[gl_capabilities_hook:FINISHED],[player_render_hook:FINISHED] Class transformer null safety: all safe AE2 Version: beta rv3-beta-6 for Forge 10.13.4.1448 CoFHCore: -[1.7.10]3.1.4-329 Mantle Environment: DO NOT REPORT THIS CRASH! Unsupported mods in environment: optifine ThermalFoundation: -[1.7.10]1.2.6-118 ThermalExpansion: -[1.7.10]4.1.5-248 TConstruct Environment: Environment healthy. ThermalDynamics: -[1.7.10]1.2.1-172 Chisel: Errors like "[FML]: Unable to lookup ..." are NOT the cause of this crash. You can safely ignore these errors. And update forge while you're at it. EnderIO: Found the following problem(s) with your installation: * Optifine is installed. This is NOT supported. * An unknown AE2 API is installed (rv3 from appliedenergistics2-rv3-beta-6.jar). Ender IO was build against API version rv2 and may or may not work with a newer version. * The RF API that is being used (1.7.10R1.3.1 from <unknown>) differes from that that is reported as being loaded (1.7.10R1.0.2 from DragonAPI 1.7.10 V18b.jar). It is a supported version, but that difference may lead to problems. This may have caused the error. Try reproducing the crash WITHOUT this/these mod(s) before reporting it. Stencil buffer state: Function set: GL30, pool: forge, bits: 8 Forestry : Warning: You have mods that change the behavior of Minecraft, ForgeModLoader, and/or Minecraft Forge to your client: Optifine These may have caused this error, and may not be supported. Try reproducing the crash WITHOUT these mods, and report it then. AE2 Integration: IC2:OFF, RotaryCraft:OFF, RC:ON, BuildCraftCore:ON, BuildCraftTransport:ON, BuildCraftBuilder:ON, RF:ON, RFItem:ON, MFR:OFF, DSU:ON, FZ:OFF, FMP:ON, RB:OFF, CLApi:OFF, Waila:ON, InvTweaks:ON, NEI:ON, CraftGuide:OFF, Mekanism:ON, ImmibisMicroblocks:OFF, BetterStorage:OFF, OpenComputers:OFF, PneumaticCraft:OFF Profiler Position: N/A (disabled) Vec3 Pool Size: 0 (0 bytes; 0 MB) allocated, 0 (0 bytes; 0 MB) used Player Count: 1 / 8; [GCEntityPlayerMP['NAMEREMOVED'/25930, l='moontest', x=193.50, y=900.00, z=155.50]] Type: Integrated Server (map_client.txt) Is Modded: Definitely; Client brand changed to 'fml,forge' OptiFine Version: OptiFine_1.7.10_HD_U_D8 Render Distance Chunks: 16 Mipmaps: 4 Anisotropic Filtering: 1 Antialiasing: 0 Multitexture: false Shaders: null OpenGlVersion: 4.5.13474 Compatibility Profile Context 22.19.162.4 OpenGlRenderer: AMD Radeon R9 200 Series OpenGlVendor: ATI Technologies Inc. CpuCount: 4 Happy to provide any more info if I can.
process
crash on changing dimensions in a rocket galacticraft galaxy space forge minecraft crash report on the bright side i bought you a teddy bear time description exception generating new chunk java lang nullpointerexception exception generating new chunk at galaxyspace solarsystem core world gen chunkproviderspacecraters replaceblocksforbiome unknown source at galaxyspace solarsystem core world gen chunkproviderspacecraters func d unknown source at net minecraft world gen chunkproviderserver originalloadchunk chunkproviderserver java at net minecraft world gen chunkproviderserver loadchunk chunkproviderserver java at net minecraft world gen chunkproviderserver func c chunkproviderserver java at net minecraft world gen chunkproviderserver func d chunkproviderserver java at net minecraft world world func e world java at net minecraft world spawneranimals func a spawneranimals java at net minecraft world worldserver func b worldserver java at net minecraft server minecraftserver func q minecraftserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server integrated integratedserver func p integratedserver java at net minecraft server minecraftserver run minecraftserver java at net minecraft server minecraftserver run minecraftserver java a detailed walkthrough of the error its code path and all known details is as follows head stacktrace at galaxyspace solarsystem core world gen chunkproviderspacecraters replaceblocksforbiome unknown source at galaxyspace solarsystem core world gen chunkproviderspacecraters func d unknown source chunk to be generated details location position hash generator randomlevelsource stacktrace at net minecraft world gen chunkproviderserver originalloadchunk chunkproviderserver java at net minecraft world gen chunkproviderserver loadchunk chunkproviderserver java at net minecraft world gen chunkproviderserver func c chunkproviderserver java at net minecraft world gen chunkproviderserver func d chunkproviderserver java at net minecraft world world func e world java at net minecraft world spawneranimals func a spawneranimals java at net minecraft world worldserver func b worldserver java affected level details level name moontest all players total chunk stats serverchunkcache drop level seed level generator id default ver features enabled false level generator options level spawn location world chunk at in contains blocks to region contains chunks to blocks to level time game time day time level dimension level storage version unknown level weather rain time now false thunder time now false level game mode error nullpointerexception null stacktrace at net minecraft server minecraftserver func q minecraftserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server integrated integratedserver func p integratedserver java at net minecraft server minecraftserver run minecraftserver java at net minecraft server minecraftserver run minecraftserver java system details details minecraft version operating system windows version java version oracle corporation java vm version java hotspot tm bit server vm mixed mode oracle corporation memory bytes mb bytes mb up to bytes mb jvm flags total xx heapdumppath mojangtricksinteldriversforperformance javaw exe minecraft exe heapdump xx xx maxgcpausemillis aabb pool size bytes mb allocated bytes mb used intcache cache tcache allocated tallocated fml mcp fml minecraft forge optifine optifine hd u mods loaded mods active states u unloaded l loaded c constructed h pre initialized i initialized j post initialized a available d disabled e errored uchijaaaaaaaaa mcp minecraft jar uchijaaaaaaaaa fml modpack jar uchijaaaaaaaaa forge modpack jar uchijaaaaaaaaa core beta minecraft jar uchijaaaaaaaaa codechickencore minecraft jar uchijaaaaaaaaa micdoodlecore minecraft jar uchijaaaaaaaaa notenoughitems notenoughitems universal jar uchijaaaaaaaaa openmodscore minecraft jar uchijaaaaaaaaa minecraft jar uchijaaaaaaaaa minecraft jar uchijaaaaaaaaa beta beta jar uchijaaaaaaaaa baubles baubles jar uchijaaaaaaaaa bibliocraft bibliocraft jar uchijaaaaaaaaa botania botania jar uchijaaaaaaaaa cofhcore cofhcore jar uchijaaaaaaaaa buildcraft core buildcraft jar uchijaaaaaaaaa buildcraft transport buildcraft jar uchijaaaaaaaaa buildcraft factory buildcraft jar uchijaaaaaaaaa buildcraft silicon buildcraft jar uchijaaaaaaaaa buildcraft robotics buildcraft jar uchijaaaaaaaaa buildcraft energy buildcraft jar uchijaaaaaaaaa buildcraft builders buildcraft jar uchijaaaaaaaaa railcraft railcraft jar uchijaaaaaaaaa twilightforest twilightforest jar uchijaaaaaaaaa forgemultipart forgemultipart universal jar uchijaaaaaaaaa chisel chisel jar uchijaaaaaaaaa carpentersblocks carpenter s blocks mc jar uchijaaaaaaaaa hardcoreenderexpansion hardcoreenderexpansion mc jar uchijaaaaaaaaa mantle mantle jar uchijaaaaaaaaa natura natura jar uchijaaaaaaaaa forestry forestry jar uchijaaaaaaaaa mrtjpcoremod mrtjpcore universal jar uchijaaaaaaaaa projred core projectred base jar uchijaaaaaaaaa thermalfoundation thermalfoundation jar uchijaaaaaaaaa extrautilities extrautilities jar uchijaaaaaaaaa thermalexpansion thermalexpansion jar uchijaaaaaaaaa waila waila jar uchijaaaaaaaaa tconstruct tconstruct jar uchijaaaaaaaaa dragonapi dragonapi jar uchijaaaaaaaaa chromaticraft chromaticraft jar uchijaaaaaaaaa damageindicatorsmod damage indicators mod jar uchijaaaaaaaaa endercore beta endercore beta jar uchijaaaaaaaaa enderio beta enderio beta jar uchijaaaaaaaaa extracells extracells jar uchijaaaaaaaaa metallurgycore metallurgycore jar uchijaaaaaaaaa metallurgy metallurgy jar uchijaaaaaaaaa extratic extratic jar uchijaaaaaaaaa galacticraftcore galacticraftcore jar uchijaaaaaaaaa galacticraftmars galacticraft planets jar uchijaaaaaaaaa galaxyspace galaxyspace stable jar uchijaaaaaaaaa geostrata geostrata jar uchijaaaaaaaaa ichunutil ichunutil jar uchijaaaaaaaaa hats hats jar uchijaaaaaaaaa inpure core inpurecore jar uchijaaaaaaaaa inventorytweaks dev inventorytweaks dev jar uchijaaaaaaaaa journeymap journeymap unlimited jar uchijaaaaaaaaa mekanism mekanism jar uchijaaaaaaaaa mekanismgenerators mekanismgenerators jar uchijaaaaaaaaa mekanismtools mekanismtools jar uchijaaaaaaaaa jar uchijaaaaaaaaa mtrm minetweakerrecipemaker jar uchijaaaaaaaaa jar uchijaaaaaaaaa morpheus morpheus jar uchijaaaaaaaaa openmods openmodslib jar uchijaaaaaaaaa openblocks openblocks jar uchijaaaaaaaaa projred integration projectred integration jar uchijaaaaaaaaa projred transmission projectred integration jar uchijaaaaaaaaa projred illumination projectred lighting jar uchijaaaaaaaaa projred exploration projectred world jar uchijaaaaaaaaa thermaldynamics thermaldynamics jar uchijaaaaaaaaa wailaharvestability wailaharvestability jar uchijaaaaaaaaa witchery witchery jar uchijaaaaaaaaa forgemicroblock forgemultipart universal jar uchijaaaaaaaaa mcmultipart forgemultipart universal jar uchijaaaaaaaaa aobd aobd jar uchijaaaaaaaaa iguanatweakstconstruct iguanatinkertweaks jar gl info error runtimeexception no opengl context found in the current thread openmodslib class transformers class transformer null safety all safe version beta beta for forge cofhcore mantle environment do not report this crash unsupported mods in environment optifine thermalfoundation thermalexpansion tconstruct environment environment healthy thermaldynamics chisel errors like unable to lookup are not the cause of this crash you can safely ignore these errors and update forge while you re at it enderio found the following problem s with your installation optifine is installed this is not supported an unknown api is installed from beta jar ender io was build against api version and may or may not work with a newer version the rf api that is being used from differes from that that is reported as being loaded from dragonapi jar it is a supported version but that difference may lead to problems this may have caused the error try reproducing the crash without this these mod s before reporting it stencil buffer state function set pool forge bits forestry warning you have mods that change the behavior of minecraft forgemodloader and or minecraft forge to your client optifine these may have caused this error and may not be supported try reproducing the crash without these mods and report it then integration off rotarycraft off rc on buildcraftcore on buildcrafttransport on buildcraftbuilder on rf on rfitem on mfr off dsu on fz off fmp on rb off clapi off waila on invtweaks on nei on craftguide off mekanism on immibismicroblocks off betterstorage off opencomputers off pneumaticcraft off profiler position n a disabled pool size bytes mb allocated bytes mb used player count type integrated server map client txt is modded definitely client brand changed to fml forge optifine version optifine hd u render distance chunks mipmaps anisotropic filtering antialiasing multitexture false shaders null openglversion compatibility profile context openglrenderer amd radeon series openglvendor ati technologies inc cpucount happy to provide any more info if i can
1
165,746
12,879,870,808
IssuesEvent
2020-07-12 01:26:50
osquery/osquery
https://api.github.com/repos/osquery/osquery
closed
Create tests for the table `sip_config`
good-first-issue macOS test
## Create tests for the table `sip_config` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/darwin/sip_config.table` Source files: - `osquery/tables/system/darwin/sip_config.cpp` Table generating function: `genSIPConfig()` Labels: `good-first-issue` `build/test`
1.0
Create tests for the table `sip_config` - ## Create tests for the table `sip_config` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/darwin/sip_config.table` Source files: - `osquery/tables/system/darwin/sip_config.cpp` Table generating function: `genSIPConfig()` Labels: `good-first-issue` `build/test`
non_process
create tests for the table sip config create tests for the table sip config create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs darwin sip config table source files osquery tables system darwin sip config cpp table generating function gensipconfig labels good first issue build test
0
1,660
4,288,774,733
IssuesEvent
2016-07-17 17:41:52
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
closed
linux preprocessor raises exception when analyzing a partition, if /etc/passwd is badly formatted
bug preprocessing
**Plaso version:** 1.3.0 dpkg -l | grep plaso ii python-plaso 1.3.0-2ppa1~trusty all Plaso Log2Timeline **Operating system Plaso is running on:** Ubuntu trusty 64b / 3.13.0-74-generic **Installation method:** ppa/GIFT **Description of problem:** Log2timeline's linux preprocessor crashes when analyzing a partition, if /etc/passwd is badly formatted. **Debug output/tracebacks:** log2timeline.py -d disk.plaso /dev/sdb1 [INFO] Data files will be loaded from /usr/share/plaso by default. Source path : /dev/sdb1 Source type : storage media image Processing started. 2016-04-25 11:12:19,256 [DEBUG] (MainProcess) PID:28115 <extraction_frontend> Starting preprocessing. Traceback (most recent call last): File "/usr/bin/log2timeline.py", line 676, in <module> if not Main(): File "/usr/bin/log2timeline.py", line 662, in Main tool.ProcessSources() File "/usr/bin/log2timeline.py", line 604, in ProcessSources timezone=self._timezone) File "/usr/lib/python2.7/dist-packages/plaso/frontend/extraction_frontend.py", line 450, in ProcessSources pre_obj = self._PreprocessSource(source_path_specs, source_type) File "/usr/lib/python2.7/dist-packages/plaso/frontend/extraction_frontend.py", line 189, in _PreprocessSource resolver_context=self._resolver_context) File "/usr/lib/python2.7/dist-packages/plaso/engine/engine.py", line 97, in PreprocessSource platform, searcher, self.knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/manager.py", line 117, in RunPlugins plugin_object.Run(searcher, knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/interface.py", line 113, in Run value = self.GetValue(searcher, knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/linux.py", line 87, in GetValue for row in reader: _csv.Error: line contains NULL byte **Source data:** Disk image.
1.0
linux preprocessor raises exception when analyzing a partition, if /etc/passwd is badly formatted - **Plaso version:** 1.3.0 dpkg -l | grep plaso ii python-plaso 1.3.0-2ppa1~trusty all Plaso Log2Timeline **Operating system Plaso is running on:** Ubuntu trusty 64b / 3.13.0-74-generic **Installation method:** ppa/GIFT **Description of problem:** Log2timeline's linux preprocessor crashes when analyzing a partition, if /etc/passwd is badly formatted. **Debug output/tracebacks:** log2timeline.py -d disk.plaso /dev/sdb1 [INFO] Data files will be loaded from /usr/share/plaso by default. Source path : /dev/sdb1 Source type : storage media image Processing started. 2016-04-25 11:12:19,256 [DEBUG] (MainProcess) PID:28115 <extraction_frontend> Starting preprocessing. Traceback (most recent call last): File "/usr/bin/log2timeline.py", line 676, in <module> if not Main(): File "/usr/bin/log2timeline.py", line 662, in Main tool.ProcessSources() File "/usr/bin/log2timeline.py", line 604, in ProcessSources timezone=self._timezone) File "/usr/lib/python2.7/dist-packages/plaso/frontend/extraction_frontend.py", line 450, in ProcessSources pre_obj = self._PreprocessSource(source_path_specs, source_type) File "/usr/lib/python2.7/dist-packages/plaso/frontend/extraction_frontend.py", line 189, in _PreprocessSource resolver_context=self._resolver_context) File "/usr/lib/python2.7/dist-packages/plaso/engine/engine.py", line 97, in PreprocessSource platform, searcher, self.knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/manager.py", line 117, in RunPlugins plugin_object.Run(searcher, knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/interface.py", line 113, in Run value = self.GetValue(searcher, knowledge_base) File "/usr/lib/python2.7/dist-packages/plaso/preprocessors/linux.py", line 87, in GetValue for row in reader: _csv.Error: line contains NULL byte **Source data:** Disk image.
process
linux preprocessor raises exception when analyzing a partition if etc passwd is badly formatted plaso version dpkg l grep plaso ii python plaso trusty all plaso operating system plaso is running on ubuntu trusty generic installation method ppa gift description of problem s linux preprocessor crashes when analyzing a partition if etc passwd is badly formatted debug output tracebacks py d disk plaso dev data files will be loaded from usr share plaso by default source path dev source type storage media image processing started mainprocess pid starting preprocessing traceback most recent call last file usr bin py line in if not main file usr bin py line in main tool processsources file usr bin py line in processsources timezone self timezone file usr lib dist packages plaso frontend extraction frontend py line in processsources pre obj self preprocesssource source path specs source type file usr lib dist packages plaso frontend extraction frontend py line in preprocesssource resolver context self resolver context file usr lib dist packages plaso engine engine py line in preprocesssource platform searcher self knowledge base file usr lib dist packages plaso preprocessors manager py line in runplugins plugin object run searcher knowledge base file usr lib dist packages plaso preprocessors interface py line in run value self getvalue searcher knowledge base file usr lib dist packages plaso preprocessors linux py line in getvalue for row in reader csv error line contains null byte source data disk image
1
17,501
2,615,145,734
IssuesEvent
2015-03-01 06:21:40
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Document gmail drag n drop out from browser -> desktop
auto-migrated Milestone-X Priority-Medium Tutorial Type-Enhancement
``` http://lists.whatwg.org/htdig.cgi/whatwg-whatwg.org/2009-August/022118.html var dragElem = document.getElementById("ID_Element_to_be_dragged"); dragElem.addEventListener( "dragstart", function(event) { event.dataTransfer.setData( "DownloadURL", "application/pdf:sample.pdf:http://example.com/example-download-data"); }, false ); ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 4 Aug 2010 at 10:33
1.0
Document gmail drag n drop out from browser -> desktop - ``` http://lists.whatwg.org/htdig.cgi/whatwg-whatwg.org/2009-August/022118.html var dragElem = document.getElementById("ID_Element_to_be_dragged"); dragElem.addEventListener( "dragstart", function(event) { event.dataTransfer.setData( "DownloadURL", "application/pdf:sample.pdf:http://example.com/example-download-data"); }, false ); ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 4 Aug 2010 at 10:33
non_process
document gmail drag n drop out from browser desktop var dragelem document getelementbyid id element to be dragged dragelem addeventlistener dragstart function event event datatransfer setdata downloadurl application pdf sample pdf false original issue reported on code google com by ericbide com on aug at
0
236,478
26,010,680,926
IssuesEvent
2022-12-21 01:10:32
hygieia/api-audit
https://api.github.com/repos/hygieia/api-audit
opened
CVE-2022-1471 (High) detected in snakeyaml-1.32.jar
security vulnerability
## CVE-2022-1471 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.32.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /repository/org/yaml/snakeyaml/1.32/snakeyaml-1.32.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.32.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hygieia/api-audit/commit/9c627a3dee72bf43b46a7cc41b8c073efba5cfab">9c627a3dee72bf43b46a7cc41b8c073efba5cfab</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p> <p>Release Date: 2022-12-01</p> <p>Fix Resolution: org.yaml:snakeyaml - 1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1471 (High) detected in snakeyaml-1.32.jar - ## CVE-2022-1471 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.32.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /repository/org/yaml/snakeyaml/1.32/snakeyaml-1.32.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.32.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hygieia/api-audit/commit/9c627a3dee72bf43b46a7cc41b8c073efba5cfab">9c627a3dee72bf43b46a7cc41b8c073efba5cfab</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p> <p>Release Date: 2022-12-01</p> <p>Fix Resolution: org.yaml:snakeyaml - 1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in snakeyaml jar cve high severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library repository org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
0
21,125
28,092,579,562
IssuesEvent
2023-03-30 13:56:37
webmachinelearning/webnn
https://api.github.com/repos/webmachinelearning/webnn
closed
Candidate Recommendation readiness tracker
process
This is a meta issue to track [Candidate Recommendation (CR)](https://www.w3.org/2021/Process-20211102/#RecsCR) publication readiness for the [Neural Network API](https://www.w3.org/TR/webnn/). ### About A Candidate Recommendation is a document that satisfies the technical requirements of the [Working Group that produced it](https://www.w3.org/groups/wg/webmachinelearning) and their dependencies, and has already received wide review. W3C publishes a Candidate Recommendation to - signal to the wider community that it is time to do a final review - gather implementation experience ### Requirements >Legend: >⚪ Not started >🟡 In progress >🟢 Completed To publish a Candidate Recommendation, in addition to meeting the [requirements for advancement](https://www.w3.org/2021/Process-20211102/#transition-reqs) a Working Group: - [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred, - 🟢 The document meets the requirements set forth in the [WG Charter](https://www.w3.org/2021/04/web-machine-learning-charter.html) and satisfies the [Use Cases](https://www.w3.org/TR/webnn/#usecases) - 🟢 The document is considered [complete and fit for purpose](https://www.w3.org/2021/Process-20211102/#ref-for-RecsCR%E2%91%A3) - 🟢 Remaining ["cr" issues](https://github.com/webmachinelearning/webnn/labels/cr) improve test coverage and alignment of core parts of the API with WebIDL conventions: - [x] https://github.com/webmachinelearning/webnn/issues/265 completed - [x] See "Done" in https://github.com/webmachinelearning/webnn/issues/210#issuecomment-1326361748 - 🟡 Post-CR enhancements: - [ ] #338 WIP - [ ] Complete https://github.com/webmachinelearning/webnn/issues/210 WIP - 🟢 [Other deliverables](https://www.w3.org/2021/04/web-machine-learning-charter.html#ig-other-deliverables): ethical considerations [documented](https://www.w3.org/TR/webmachinelearning-ethics/) - [x] must document changes to dependencies during the development of the specification, - 🟢 No new [normative references](https://www.w3.org/TR/webnn/#normative) since [FPWD](https://www.w3.org/TR/2021/WD-webnn-20210622/) - [x] must document how adequate implementation experience will be demonstrated, - 🟢 [web-platform-tests test suite](https://github.com/web-platform-tests/wpt/tree/master/webnn) / https://github.com/webmachinelearning/webnn/issues/265 / [wpt.fyi](https://wpt.fyi/results/webnn?label=experimental&label=master&aligned&view=subtest) - 🟡 Preliminary implementation reports based on one or more of: - 🟡 JS implementations: [webnn-baseline](https://github.com/webmachinelearning/webnn-baseline) (pure JS), [webnn-polyfill](https://github.com/webmachinelearning/webnn-polyfill) (use TF.js backends) - 🟡 Chromium implementation: [Intent to Prototype](https://groups.google.com/a/chromium.org/g/blink-dev/c/PD6TDMDS9mg), [source](https://source.chromium.org/chromium/chromium/src/+/main:third_party/blink/renderer/modules/ml/webnn/), [tracker bug](https://crbug.com/1273291), [issues](https://bugs.chromium.org/p/chromium/issues/list?q=component:Blink%3EWebML), [patches](https://chromium-review.googlesource.com/q/hashtag:webnn+(status:open%20OR%20status:merged)) - 🟡 Chromium backends: [XNNPACK](https://chromium-review.googlesource.com/c/chromium/src/+/3684745) and [DirectML](https://chromium-review.googlesource.com/c/chromium/src/+/3915513) ([design doc](https://docs.google.com/document/d/1TMs36IE9wL9rNuh8lriGr51S8MU1JezU2SCZy-YqJ3Y/)) - Note: per Chromium dev [best practices](https://chromium.googlesource.com/chromium/src.git/+/HEAD/docs/cl_respect.md#mind-your-reviewer) the initial prototyping happens in big patches ("big CLs") but is landed in multiple small patches ("small CLs"), see the [tracker issue](https://bugs.chromium.org/p/chromium/issues/detail?id=1273291) for the latest landed patches - [x] must specify the deadline for comments, which must be at least 28 days after publication, and should be longer for complex documents, - [x] must show that the specification has received wide review, and - 🟢 Wide review tracked in https://github.com/webmachinelearning/webnn/issues/239 - Implementation-specific reviews of interest (informational, not a complete list): - 🟢 Chromium security review for [accelerated backends](https://bugs.chromium.org/p/chromium/issues/detail?id=1381569) - [x] may identify features in the document as at risk. These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation. - 🟢 WebGPU interoperability [identified](https://www.w3.org/2022/11/03-webmachinelearning-minutes.html#t05) as a feature at risk for CR publication, needs more implementation experience, SOTD updated accordingly in #340
1.0
Candidate Recommendation readiness tracker - This is a meta issue to track [Candidate Recommendation (CR)](https://www.w3.org/2021/Process-20211102/#RecsCR) publication readiness for the [Neural Network API](https://www.w3.org/TR/webnn/). ### About A Candidate Recommendation is a document that satisfies the technical requirements of the [Working Group that produced it](https://www.w3.org/groups/wg/webmachinelearning) and their dependencies, and has already received wide review. W3C publishes a Candidate Recommendation to - signal to the wider community that it is time to do a final review - gather implementation experience ### Requirements >Legend: >⚪ Not started >🟡 In progress >🟢 Completed To publish a Candidate Recommendation, in addition to meeting the [requirements for advancement](https://www.w3.org/2021/Process-20211102/#transition-reqs) a Working Group: - [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred, - 🟢 The document meets the requirements set forth in the [WG Charter](https://www.w3.org/2021/04/web-machine-learning-charter.html) and satisfies the [Use Cases](https://www.w3.org/TR/webnn/#usecases) - 🟢 The document is considered [complete and fit for purpose](https://www.w3.org/2021/Process-20211102/#ref-for-RecsCR%E2%91%A3) - 🟢 Remaining ["cr" issues](https://github.com/webmachinelearning/webnn/labels/cr) improve test coverage and alignment of core parts of the API with WebIDL conventions: - [x] https://github.com/webmachinelearning/webnn/issues/265 completed - [x] See "Done" in https://github.com/webmachinelearning/webnn/issues/210#issuecomment-1326361748 - 🟡 Post-CR enhancements: - [ ] #338 WIP - [ ] Complete https://github.com/webmachinelearning/webnn/issues/210 WIP - 🟢 [Other deliverables](https://www.w3.org/2021/04/web-machine-learning-charter.html#ig-other-deliverables): ethical considerations [documented](https://www.w3.org/TR/webmachinelearning-ethics/) - [x] must document changes to dependencies during the development of the specification, - 🟢 No new [normative references](https://www.w3.org/TR/webnn/#normative) since [FPWD](https://www.w3.org/TR/2021/WD-webnn-20210622/) - [x] must document how adequate implementation experience will be demonstrated, - 🟢 [web-platform-tests test suite](https://github.com/web-platform-tests/wpt/tree/master/webnn) / https://github.com/webmachinelearning/webnn/issues/265 / [wpt.fyi](https://wpt.fyi/results/webnn?label=experimental&label=master&aligned&view=subtest) - 🟡 Preliminary implementation reports based on one or more of: - 🟡 JS implementations: [webnn-baseline](https://github.com/webmachinelearning/webnn-baseline) (pure JS), [webnn-polyfill](https://github.com/webmachinelearning/webnn-polyfill) (use TF.js backends) - 🟡 Chromium implementation: [Intent to Prototype](https://groups.google.com/a/chromium.org/g/blink-dev/c/PD6TDMDS9mg), [source](https://source.chromium.org/chromium/chromium/src/+/main:third_party/blink/renderer/modules/ml/webnn/), [tracker bug](https://crbug.com/1273291), [issues](https://bugs.chromium.org/p/chromium/issues/list?q=component:Blink%3EWebML), [patches](https://chromium-review.googlesource.com/q/hashtag:webnn+(status:open%20OR%20status:merged)) - 🟡 Chromium backends: [XNNPACK](https://chromium-review.googlesource.com/c/chromium/src/+/3684745) and [DirectML](https://chromium-review.googlesource.com/c/chromium/src/+/3915513) ([design doc](https://docs.google.com/document/d/1TMs36IE9wL9rNuh8lriGr51S8MU1JezU2SCZy-YqJ3Y/)) - Note: per Chromium dev [best practices](https://chromium.googlesource.com/chromium/src.git/+/HEAD/docs/cl_respect.md#mind-your-reviewer) the initial prototyping happens in big patches ("big CLs") but is landed in multiple small patches ("small CLs"), see the [tracker issue](https://bugs.chromium.org/p/chromium/issues/detail?id=1273291) for the latest landed patches - [x] must specify the deadline for comments, which must be at least 28 days after publication, and should be longer for complex documents, - [x] must show that the specification has received wide review, and - 🟢 Wide review tracked in https://github.com/webmachinelearning/webnn/issues/239 - Implementation-specific reviews of interest (informational, not a complete list): - 🟢 Chromium security review for [accelerated backends](https://bugs.chromium.org/p/chromium/issues/detail?id=1381569) - [x] may identify features in the document as at risk. These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation. - 🟢 WebGPU interoperability [identified](https://www.w3.org/2022/11/03-webmachinelearning-minutes.html#t05) as a feature at risk for CR publication, needs more implementation experience, SOTD updated accordingly in #340
process
candidate recommendation readiness tracker this is a meta issue to track publication readiness for the about a candidate recommendation is a document that satisfies the technical requirements of the and their dependencies and has already received wide review publishes a candidate recommendation to signal to the wider community that it is time to do a final review gather implementation experience requirements legend ⚪ not started 🟡 in progress 🟢 completed to publish a candidate recommendation in addition to meeting the a working group must show that the specification has met all working group requirements or explain why the requirements have changed or been deferred 🟢 the document meets the requirements set forth in the and satisfies the 🟢 the document is considered 🟢 remaining improve test coverage and alignment of core parts of the api with webidl conventions completed see done in 🟡 post cr enhancements wip complete wip 🟢 ethical considerations must document changes to dependencies during the development of the specification 🟢 no new since must document how adequate implementation experience will be demonstrated 🟢 🟡 preliminary implementation reports based on one or more of 🟡 js implementations pure js use tf js backends 🟡 chromium implementation 🟡 chromium backends and note per chromium dev the initial prototyping happens in big patches big cls but is landed in multiple small patches small cls see the for the latest landed patches must specify the deadline for comments which must be at least days after publication and should be longer for complex documents must show that the specification has received wide review and 🟢 wide review tracked in implementation specific reviews of interest informational not a complete list 🟢 chromium security review for may identify features in the document as at risk these features may be removed before advancement to proposed recommendation without a requirement to publish a new candidate recommendation 🟢 webgpu interoperability as a feature at risk for cr publication needs more implementation experience sotd updated accordingly in
1
333,852
29,813,078,379
IssuesEvent
2023-06-16 16:36:58
bcgov/SIMS
https://api.github.com/repos/bcgov/SIMS
closed
e2e - Basic tests for the low coverage parts of the application
E2E/Unit tests
Create basic e2e tests for - Institution Account Creation - Designation Agreement - Program - Offferings
1.0
e2e - Basic tests for the low coverage parts of the application - Create basic e2e tests for - Institution Account Creation - Designation Agreement - Program - Offferings
non_process
basic tests for the low coverage parts of the application create basic tests for institution account creation designation agreement program offferings
0
247,885
26,749,121,081
IssuesEvent
2023-01-30 18:07:41
istio/istio
https://api.github.com/repos/istio/istio
closed
2022 Audit related fixes
area/product security
This issue is historical, just to create references to PRs that previously did not have an associated issue Issue number | Description | Related PRs -- | -- | -- 1 | Possible disk exhaustion | https://github.com/istio/istio/pull/41705 2 | Arbitrary file write during archive extraction | https://github.com/istio/istio/pull/41786 3 | File not closed | https://github.com/istio/istio/pull/41786 4 | Len of new byte slice controlled by potentially untrusted file size | https://github.com/istio/istio/pull/41894 5 | Possible memory exhaustions through io.ReadAll | https://github.com/istio/istio/pull/41894 6 | Use of md5 | https://github.com/istio/istio/pull/41930 7 | istio uses insecure math/rand instead of crypto/rand | https://github.com/istio/istio/pull/41930 8 | Istio skips certificate verification with InsecureSkipVerify: true | https://github.com/istio/istio/pull/41930 9 | Unhandled errors | https://github.com/istio/istio/pull/41902 10 | Use of deprecated 3rd party library | https://github.com/istio/istio/pull/41343 11 | TOCTOU race conditions | https://github.com/istio/istio/pull/42040 12 | H2c handlers are uncapped | https://github.com/istio/istio/pull/41872 13 | STS Service server is susceptible to DoS if debug mode is enabled | https://github.com/istio/istio/pull/41962
True
2022 Audit related fixes - This issue is historical, just to create references to PRs that previously did not have an associated issue Issue number | Description | Related PRs -- | -- | -- 1 | Possible disk exhaustion | https://github.com/istio/istio/pull/41705 2 | Arbitrary file write during archive extraction | https://github.com/istio/istio/pull/41786 3 | File not closed | https://github.com/istio/istio/pull/41786 4 | Len of new byte slice controlled by potentially untrusted file size | https://github.com/istio/istio/pull/41894 5 | Possible memory exhaustions through io.ReadAll | https://github.com/istio/istio/pull/41894 6 | Use of md5 | https://github.com/istio/istio/pull/41930 7 | istio uses insecure math/rand instead of crypto/rand | https://github.com/istio/istio/pull/41930 8 | Istio skips certificate verification with InsecureSkipVerify: true | https://github.com/istio/istio/pull/41930 9 | Unhandled errors | https://github.com/istio/istio/pull/41902 10 | Use of deprecated 3rd party library | https://github.com/istio/istio/pull/41343 11 | TOCTOU race conditions | https://github.com/istio/istio/pull/42040 12 | H2c handlers are uncapped | https://github.com/istio/istio/pull/41872 13 | STS Service server is susceptible to DoS if debug mode is enabled | https://github.com/istio/istio/pull/41962
non_process
audit related fixes this issue is historical just to create references to prs that previously did not have an associated issue issue number description related prs possible disk exhaustion arbitrary file write during archive extraction file not closed len of new byte slice controlled by potentially untrusted file size possible memory exhaustions through io readall use of istio uses insecure math rand instead of crypto rand istio skips certificate verification with insecureskipverify true unhandled errors use of deprecated party library toctou race conditions handlers are uncapped sts service server is susceptible to dos if debug mode is enabled
0
16,198
20,691,295,572
IssuesEvent
2022-03-11 00:39:32
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Need ability to break down VISITORS data by minute
enhancement log-processing command-line options
First off, Thanks so much for such an awesome and amazing application! Currently I can get the visitors data break down by day or by hour, but would really like to be able to get it by minute as well. At the very least I am looking for every 5 minutes, but I know that many people might need it by minute. It would be awesome to just add an option to `--date-spec` for `min` then just append the minute to the data like how the hour is added when you do `--date-spec=hr`. So in the end with `--date-spec=min` you would see `202202150259`, etc. I know that this could cause an issue on the front end GUI as there would be a lot of data to show, but I only need this for the CLI as I just need the JSON data for this. So maybe as a first step this can be added only for the cli and if more requests come from it then you can add it to the GUI later on. Thanks
1.0
Need ability to break down VISITORS data by minute - First off, Thanks so much for such an awesome and amazing application! Currently I can get the visitors data break down by day or by hour, but would really like to be able to get it by minute as well. At the very least I am looking for every 5 minutes, but I know that many people might need it by minute. It would be awesome to just add an option to `--date-spec` for `min` then just append the minute to the data like how the hour is added when you do `--date-spec=hr`. So in the end with `--date-spec=min` you would see `202202150259`, etc. I know that this could cause an issue on the front end GUI as there would be a lot of data to show, but I only need this for the CLI as I just need the JSON data for this. So maybe as a first step this can be added only for the cli and if more requests come from it then you can add it to the GUI later on. Thanks
process
need ability to break down visitors data by minute first off thanks so much for such an awesome and amazing application currently i can get the visitors data break down by day or by hour but would really like to be able to get it by minute as well at the very least i am looking for every minutes but i know that many people might need it by minute it would be awesome to just add an option to date spec for min then just append the minute to the data like how the hour is added when you do date spec hr so in the end with date spec min you would see etc i know that this could cause an issue on the front end gui as there would be a lot of data to show but i only need this for the cli as i just need the json data for this so maybe as a first step this can be added only for the cli and if more requests come from it then you can add it to the gui later on thanks
1
28,152
4,366,133,273
IssuesEvent
2016-08-03 13:28:43
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Unable to deploy database in e2e test: secret never got mounted
component/storage kind/test-flake priority/P0
Post Rebase (kube 1.3.0) on our RHEL 7.2 CI machines some deployments are failing because they are unable to mount a secret occasionally. https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_integration/3091/ https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_integration/3091/s3/download/test-end-to-end-docker/logs/container-origin.log Opened kubernetes/kubernetes#28750
1.0
Unable to deploy database in e2e test: secret never got mounted - Post Rebase (kube 1.3.0) on our RHEL 7.2 CI machines some deployments are failing because they are unable to mount a secret occasionally. https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_integration/3091/ https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_integration/3091/s3/download/test-end-to-end-docker/logs/container-origin.log Opened kubernetes/kubernetes#28750
non_process
unable to deploy database in test secret never got mounted post rebase kube on our rhel ci machines some deployments are failing because they are unable to mount a secret occasionally opened kubernetes kubernetes
0
273,818
20,817,162,961
IssuesEvent
2022-03-18 11:38:00
T4rikA/lively.next
https://api.github.com/repos/T4rikA/lively.next
opened
Project Submitting
documentation
As a software engineering student, I'd like to have all the things ready that I need for submitting this project --- Other things that need to be submitted: - [ ] Screencast (narrated or with subtitles - [ ] Abstract (txt) - [ ] Figure / Screenshot - [ ] Sourcecode (MIT License) - [ ] Documentation - [ ] README.md --- ## Conditions of satisfaction - [ ] Should have a UI (#36) - [ ] Should have working merging of morphs - [ ] Should have working world merging - [ ] All of the other things are ready and in a zip file
1.0
Project Submitting - As a software engineering student, I'd like to have all the things ready that I need for submitting this project --- Other things that need to be submitted: - [ ] Screencast (narrated or with subtitles - [ ] Abstract (txt) - [ ] Figure / Screenshot - [ ] Sourcecode (MIT License) - [ ] Documentation - [ ] README.md --- ## Conditions of satisfaction - [ ] Should have a UI (#36) - [ ] Should have working merging of morphs - [ ] Should have working world merging - [ ] All of the other things are ready and in a zip file
non_process
project submitting as a software engineering student i d like to have all the things ready that i need for submitting this project other things that need to be submitted screencast narrated or with subtitles abstract txt figure screenshot sourcecode mit license documentation readme md conditions of satisfaction should have a ui should have working merging of morphs should have working world merging all of the other things are ready and in a zip file
0
86,177
10,724,466,768
IssuesEvent
2019-10-28 01:48:11
RoboJackets/igvc-software
https://api.github.com/repos/RoboJackets/igvc-software
opened
Integrate SLAM into navigation stack
area ➤ localization area ➤ mapping level ➤ hard type ➤ design document type ➤ new feature type ➤ research
One of IGVC's biggest goals has been implementing SLAM for our codebase. This year, let's try to get this to work.
1.0
Integrate SLAM into navigation stack - One of IGVC's biggest goals has been implementing SLAM for our codebase. This year, let's try to get this to work.
non_process
integrate slam into navigation stack one of igvc s biggest goals has been implementing slam for our codebase this year let s try to get this to work
0
70,768
18,270,706,716
IssuesEvent
2021-10-04 13:33:47
reactjs/it.reactjs.org
https://api.github.com/repos/reactjs/it.reactjs.org
closed
Translation for the page 'Advanced Guides -> Forwarding Refs'
:building_construction: Work in progress
# Please help us translating this page in Italian! **Aiutaci a tradurre questa pagina in Italiano!** You can find all the details about how to help here: https://github.com/reactjs/it.reactjs.org/issues/1 ## GRAZIE! :smile:
1.0
Translation for the page 'Advanced Guides -> Forwarding Refs' - # Please help us translating this page in Italian! **Aiutaci a tradurre questa pagina in Italiano!** You can find all the details about how to help here: https://github.com/reactjs/it.reactjs.org/issues/1 ## GRAZIE! :smile:
non_process
translation for the page advanced guides forwarding refs please help us translating this page in italian aiutaci a tradurre questa pagina in italiano you can find all the details about how to help here grazie smile
0
6,543
9,634,832,612
IssuesEvent
2019-05-15 22:28:14
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
Run docs, samples, system, lint tests on node 10
type: process
Our kokoro config is currently set up to run all these tests on node.js 8. That's going EOL by the EOY, and it's slower than 10. Let's upgrade!
1.0
Run docs, samples, system, lint tests on node 10 - Our kokoro config is currently set up to run all these tests on node.js 8. That's going EOL by the EOY, and it's slower than 10. Let's upgrade!
process
run docs samples system lint tests on node our kokoro config is currently set up to run all these tests on node js that s going eol by the eoy and it s slower than let s upgrade
1
8,285
3,152,600,839
IssuesEvent
2015-09-16 14:33:52
mgibbs189/custom-field-suite
https://api.github.com/repos/mgibbs189/custom-field-suite
closed
Tab field type is undocumented
documentation
As in title. It's not complicated, and I figured it out(it's very handy), but there should still be proper docs.
1.0
Tab field type is undocumented - As in title. It's not complicated, and I figured it out(it's very handy), but there should still be proper docs.
non_process
tab field type is undocumented as in title it s not complicated and i figured it out it s very handy but there should still be proper docs
0
441,865
12,733,751,303
IssuesEvent
2020-06-25 12:50:59
hochschule-darmstadt/openartbrowser
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
opened
Empty Tool-tipps
bug medium priority small effort
**Describe the bug** Some tool-tips are empty when there is no description for the hovered item **To Reproduce** Steps to reproduce the behavior: 1. Go to https://openartbrowser.org/de/artist/Q7814 2. Click 'more' 3. Hover over "Italienische Renaissance Gemälde" **Expected behavior** No tool-tip should be shown **Screenshots** ![image](https://user-images.githubusercontent.com/28682931/85722736-e0ea1700-b6f2-11ea-8f7c-9c2ecfec1443.png)
1.0
Empty Tool-tipps - **Describe the bug** Some tool-tips are empty when there is no description for the hovered item **To Reproduce** Steps to reproduce the behavior: 1. Go to https://openartbrowser.org/de/artist/Q7814 2. Click 'more' 3. Hover over "Italienische Renaissance Gemälde" **Expected behavior** No tool-tip should be shown **Screenshots** ![image](https://user-images.githubusercontent.com/28682931/85722736-e0ea1700-b6f2-11ea-8f7c-9c2ecfec1443.png)
non_process
empty tool tipps describe the bug some tool tips are empty when there is no description for the hovered item to reproduce steps to reproduce the behavior go to click more hover over italienische renaissance gemälde expected behavior no tool tip should be shown screenshots
0
18,376
24,503,562,501
IssuesEvent
2022-10-10 14:34:52
fadeoutsoftware/WASDI
https://api.github.com/repos/fadeoutsoftware/WASDI
opened
Docker > Rename images builded
enhancement P3 app / processor
When we have to redeploy containers in WASDI it is not so easy. ONE of the "difficulties" is: - how to know what is a python application if I need to redeploy Python applications? - how to know what is a conda app if I need to redeploy Conda apps? - etc Today containers are named "wasdi/<something>". Example: ``` wasdi/s2rgb ``` Could we imagine to rename containers? Idea 1: - name: <template name>/<something> - examples: ``` python37/myApp1 ``` ``` conda/myApp2 ``` Idea 2: - name: wasdi/<template name>/<something> - examples: ``` wasdi/python37/myApp1 ``` ``` wasdi/conda/myApp2 ``` Of course it is just a cold idea: we need to analyze limitations (name length, characters we can use, etc)
1.0
Docker > Rename images builded - When we have to redeploy containers in WASDI it is not so easy. ONE of the "difficulties" is: - how to know what is a python application if I need to redeploy Python applications? - how to know what is a conda app if I need to redeploy Conda apps? - etc Today containers are named "wasdi/<something>". Example: ``` wasdi/s2rgb ``` Could we imagine to rename containers? Idea 1: - name: <template name>/<something> - examples: ``` python37/myApp1 ``` ``` conda/myApp2 ``` Idea 2: - name: wasdi/<template name>/<something> - examples: ``` wasdi/python37/myApp1 ``` ``` wasdi/conda/myApp2 ``` Of course it is just a cold idea: we need to analyze limitations (name length, characters we can use, etc)
process
docker rename images builded when we have to redeploy containers in wasdi it is not so easy one of the difficulties is how to know what is a python application if i need to redeploy python applications how to know what is a conda app if i need to redeploy conda apps etc today containers are named wasdi example wasdi could we imagine to rename containers idea name examples conda idea name wasdi examples wasdi wasdi conda of course it is just a cold idea we need to analyze limitations name length characters we can use etc
1
14,199
17,100,068,851
IssuesEvent
2021-07-09 09:56:41
googleapis/python-test-utils
https://api.github.com/repos/googleapis/python-test-utils
opened
Add tests for maybe_fail_import()
type: process
Context: https://github.com/googleapis/google-cloud-python/pull/8840#issuecomment-876524337 Refactoring the utility required a bit more work than it ideally should, thus we should cover it with tests.
1.0
Add tests for maybe_fail_import() - Context: https://github.com/googleapis/google-cloud-python/pull/8840#issuecomment-876524337 Refactoring the utility required a bit more work than it ideally should, thus we should cover it with tests.
process
add tests for maybe fail import context refactoring the utility required a bit more work than it ideally should thus we should cover it with tests
1
97,714
20,378,411,661
IssuesEvent
2022-02-21 18:07:38
jvegax/Acme-One
https://api.github.com/repos/jvegax/Acme-One
closed
Task-014 : Internationalise the system
code 🧑‍💻
The system must be internationalised in English and Spanish. Other mainstream languages are welcome, but not required.
1.0
Task-014 : Internationalise the system - The system must be internationalised in English and Spanish. Other mainstream languages are welcome, but not required.
non_process
task internationalise the system the system must be internationalised in english and spanish other mainstream languages are welcome but not required
0
213,813
24,022,401,028
IssuesEvent
2022-09-15 08:45:02
sast-automation-dev/openmrs-core-25
https://api.github.com/repos/sast-automation-dev/openmrs-core-25
opened
jquery-1.7.1.min.js: 5 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jquery - 3.5.0 | &#10060; | | [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - v3.0.0 | &#10060; | | [CVE-2020-7656](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jquery - 1.9.0 | &#10060; | | [CVE-2012-6708](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - v1.9.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-7656</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
True
jquery-1.7.1.min.js: 5 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jquery - 3.5.0 | &#10060; | | [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - v3.0.0 | &#10060; | | [CVE-2020-7656](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jquery - 1.9.0 | &#10060; | | [CVE-2012-6708](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.7.1.min.js | Direct | jQuery - v1.9.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-7656</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>jquery-1.7.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to vulnerable library: /webapp/src/main/webapp/WEB-INF/view/scripts/jquery/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-25/commit/93ed58729dca2e5b94957722a7c8d1f8c02c687b">93ed58729dca2e5b94957722a7c8d1f8c02c687b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
non_process
jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium jquery min js direct jquery medium jquery min js direct jquery medium jquery min js direct jquery medium jquery min js direct jquery medium jquery min js direct jquery details cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library webapp src main webapp web inf view scripts jquery jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
59,042
11,940,552,065
IssuesEvent
2020-04-02 16:54:24
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Code Snippet react component has a strange gradient also apparent on hover
component: code-snippet type: bug 🐛
Carbon 10.10.3 Here's a screen shot: <img width="772" alt="Screen Shot 2020-04-02 at 18 59 39" src="https://user-images.githubusercontent.com/7255583/78272574-5ab89d00-7516-11ea-8707-4d4ad26fbba4.png">
1.0
Code Snippet react component has a strange gradient also apparent on hover - Carbon 10.10.3 Here's a screen shot: <img width="772" alt="Screen Shot 2020-04-02 at 18 59 39" src="https://user-images.githubusercontent.com/7255583/78272574-5ab89d00-7516-11ea-8707-4d4ad26fbba4.png">
non_process
code snippet react component has a strange gradient also apparent on hover carbon here s a screen shot img width alt screen shot at src
0
38,315
5,173,562,276
IssuesEvent
2017-01-18 16:22:47
ngageoint/hootenanny-ui
https://api.github.com/repos/ngageoint/hootenanny-ui
closed
Issue with Cookie Cutter Conflation
Category: Test Identified During Regression Test Status: Ready for Test Type: Bug
Attempted the following method: **Cookie Cutter & Horizontal** For this example we’ll need to create two custom translations, one for the DC Street Centerline Data* described in and a second simple translation to ensure that the OSM highway data for DC maintains the correct osm tags. Ingest DC Street datasets using the recently created custom translation files. Note the different Translation Schema files used to import each dataset. - district_of_columbia_highway.zip - Street_Centerline_Light.shp Return to Map and select the Street Centerlines Light as the Reference Dataset, dc highway osm as the Secondary dataset. Click ‘Conflate’ Change the value for Type to Cookie Cutter & Horizontal. Hit Conflate. Note the conflation time will vary depending on the specs of the machine. This example took about 10-15 min to run locally. 50+ reviews should appear. ----> Instead of launching into review mode, it kicks back an error (will be happy to include the log upon request due to length) Hypothesized it could be any issue with the data, but I attempted the above cookie cut conflation on Hoot NOME with zero issues. Would somebody mind trying to replicate the results on hoot release?
3.0
Issue with Cookie Cutter Conflation - Attempted the following method: **Cookie Cutter & Horizontal** For this example we’ll need to create two custom translations, one for the DC Street Centerline Data* described in and a second simple translation to ensure that the OSM highway data for DC maintains the correct osm tags. Ingest DC Street datasets using the recently created custom translation files. Note the different Translation Schema files used to import each dataset. - district_of_columbia_highway.zip - Street_Centerline_Light.shp Return to Map and select the Street Centerlines Light as the Reference Dataset, dc highway osm as the Secondary dataset. Click ‘Conflate’ Change the value for Type to Cookie Cutter & Horizontal. Hit Conflate. Note the conflation time will vary depending on the specs of the machine. This example took about 10-15 min to run locally. 50+ reviews should appear. ----> Instead of launching into review mode, it kicks back an error (will be happy to include the log upon request due to length) Hypothesized it could be any issue with the data, but I attempted the above cookie cut conflation on Hoot NOME with zero issues. Would somebody mind trying to replicate the results on hoot release?
non_process
issue with cookie cutter conflation attempted the following method cookie cutter horizontal for this example we’ll need to create two custom translations one for the dc street centerline data described in and a second simple translation to ensure that the osm highway data for dc maintains the correct osm tags ingest dc street datasets using the recently created custom translation files note the different translation schema files used to import each dataset district of columbia highway zip street centerline light shp return to map and select the street centerlines light as the reference dataset dc highway osm as the secondary dataset click ‘conflate’ change the value for type to cookie cutter horizontal hit conflate note the conflation time will vary depending on the specs of the machine this example took about min to run locally reviews should appear instead of launching into review mode it kicks back an error will be happy to include the log upon request due to length hypothesized it could be any issue with the data but i attempted the above cookie cut conflation on hoot nome with zero issues would somebody mind trying to replicate the results on hoot release
0
491,422
14,163,817,547
IssuesEvent
2020-11-12 03:20:32
woocommerce/woocommerce-admin
https://api.github.com/repos/woocommerce/woocommerce-admin
closed
The wp-admin theme color is not used on wc-admin pages
[Priority] High [Type] Bug
**Describe the bug** On latest `main` of wc-admin something is wrong with use of the wp-admin theme color CSS var, it appears that it is not set and it makes a lot of elements look visually defective or invisible. **To Reproduce** 1. Install WooCommerce 4.6.0 2. Install wc-admin from your local dev environment 3. Visit the profile wizard 4. Note that continue buttons appear to be missing **Expected behavior** I expect to see the default theme color as the background color making buttons and other themed accents visible. **Screenshots** <img width="675" alt="Screenshot 2020-10-16 at 3 24 56 PM" src="https://user-images.githubusercontent.com/1281828/96206032-7d973400-0fc4-11eb-9d2f-6b1cf22d8daf.png"> <img width="604" alt="Screenshot 2020-10-16 at 3 25 07 PM" src="https://user-images.githubusercontent.com/1281828/96206035-80922480-0fc4-11eb-97fd-50ec42d6217d.png"> **Desktop (please complete the following information):** * OS: Mac OS * Browser Chrome 86
1.0
The wp-admin theme color is not used on wc-admin pages - **Describe the bug** On latest `main` of wc-admin something is wrong with use of the wp-admin theme color CSS var, it appears that it is not set and it makes a lot of elements look visually defective or invisible. **To Reproduce** 1. Install WooCommerce 4.6.0 2. Install wc-admin from your local dev environment 3. Visit the profile wizard 4. Note that continue buttons appear to be missing **Expected behavior** I expect to see the default theme color as the background color making buttons and other themed accents visible. **Screenshots** <img width="675" alt="Screenshot 2020-10-16 at 3 24 56 PM" src="https://user-images.githubusercontent.com/1281828/96206032-7d973400-0fc4-11eb-9d2f-6b1cf22d8daf.png"> <img width="604" alt="Screenshot 2020-10-16 at 3 25 07 PM" src="https://user-images.githubusercontent.com/1281828/96206035-80922480-0fc4-11eb-97fd-50ec42d6217d.png"> **Desktop (please complete the following information):** * OS: Mac OS * Browser Chrome 86
non_process
the wp admin theme color is not used on wc admin pages describe the bug on latest main of wc admin something is wrong with use of the wp admin theme color css var it appears that it is not set and it makes a lot of elements look visually defective or invisible to reproduce install woocommerce install wc admin from your local dev environment visit the profile wizard note that continue buttons appear to be missing expected behavior i expect to see the default theme color as the background color making buttons and other themed accents visible screenshots img width alt screenshot at pm src img width alt screenshot at pm src desktop please complete the following information os mac os browser chrome
0