Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
22
2,496,263,003
IssuesEvent
2015-01-06 18:14:42
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
Cell Potency
biological_process imported
_From [rgar...@eagle-i.org](https://code.google.com/u/111247205719752845822/) on March 25, 2013 08:50:35_ \<b>**** Use the form below to request a new term ****</b> \<b>**** Scroll down to see a term request example ****</b> &#13; \<b>Please indicate the label for the proposed term:</b> Cell Potency&#13; &#13; \<b>Please provide a textual definition (with source):</b> The differentiation potential of a cell ( \<a href="https://en.wikipedia.org/wiki/Cell_potency" rel="nofollow">https://en.wikipedia.org/wiki/Cell_potency</a> )&#13; &#13; \<b>Please add an example of usage for proposed term:</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[ ] Instrument</b> [X] Biological process&#13; \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info:</b> &#13; &#13; &#13; &#13; \<b>*** Term request example ****</b> &#13; \<b>Please indicate the label for the proposed term: four-terminal resistance</b> \<b>sensor</b> &#13; &#13; Please provide a textual definition (with source): "Four-terminal&#13; \<b>resistance sensors are electrical impedance measuring instruments that use</b> \<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b> \<b>accurate measurements that can be used to compute a material's electrical</b> resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>&#13; &#13; &#13; \<b>Please add an example of usage for proposed term: Measuring the inherent</b> \<b>(per square) resistance of doped silicon.</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[X] Instrument</b> \<b>[ ] Biological process</b> \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b> &#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=203_
1.0
Cell Potency - _From [rgar...@eagle-i.org](https://code.google.com/u/111247205719752845822/) on March 25, 2013 08:50:35_ \<b>**** Use the form below to request a new term ****</b> \<b>**** Scroll down to see a term request example ****</b> &#13; \<b>Please indicate the label for the proposed term:</b> Cell Potency&#13; &#13; \<b>Please provide a textual definition (with source):</b> The differentiation potential of a cell ( \<a href="https://en.wikipedia.org/wiki/Cell_potency" rel="nofollow">https://en.wikipedia.org/wiki/Cell_potency</a> )&#13; &#13; \<b>Please add an example of usage for proposed term:</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[ ] Instrument</b> [X] Biological process&#13; \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info:</b> &#13; &#13; &#13; &#13; \<b>*** Term request example ****</b> &#13; \<b>Please indicate the label for the proposed term: four-terminal resistance</b> \<b>sensor</b> &#13; &#13; Please provide a textual definition (with source): "Four-terminal&#13; \<b>resistance sensors are electrical impedance measuring instruments that use</b> \<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b> \<b>accurate measurements that can be used to compute a material's electrical</b> resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>&#13; &#13; &#13; \<b>Please add an example of usage for proposed term: Measuring the inherent</b> \<b>(per square) resistance of doped silicon.</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[X] Instrument</b> \<b>[ ] Biological process</b> \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b> &#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=203_
process
cell potency from on march use the form below to request a new term scroll down to see a term request example please indicate the label for the proposed term cell potency please provide a textual definition with source the differentiation potential of a cell please add an example of usage for proposed term please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info term request example please indicate the label for the proposed term four terminal resistance sensor please provide a textual definition with source four terminal resistance sensors are electrical impedance measuring instruments that use separate pairs of current carrying and voltage sensing electrodes to make accurate measurements that can be used to compute a material s electrical resistance please add an example of usage for proposed term measuring the inherent per square resistance of doped silicon please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info aka sensors wire sensor or point probe original issue
1
21,390
29,202,231,792
IssuesEvent
2023-05-21 00:37:38
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Site Reliability Engineer (SRE) na Coodesh
SALVADOR PJ BANCO DE DADOS BIG DATA REDES PLENO POSTGRESQL KUBERNETES DEVOPS AWS REQUISITOS LINUX REMOTO PROCESSOS GITHUB AZURE SEGURANÇA UMA ANALYTICS ESPANHOL SISTEMAS OPERACIONAIS MACHINE LEARNING NEGÓCIOS GCP INTELIGÊNCIA ARTIFICIAL OPENSHIFT ARQUITETURA DE DADOS MONITORAMENTO SRE Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/site-reliability-engineer-sre-180605404?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Tatic</strong> está em busca de <strong><ins>Site Reliability Engineer (SRE)</ins></strong> para compor seu time!</p> <p>Buscamos <strong><em><ins>Pessoas SRE - Site Reliability Engineer</ins></em></strong><em><ins> </ins></em><strong><em><ins>Pleno</ins></em></strong> para atuar 100% remoto na Tatic.&nbsp;</p> <p><strong>Conheça a Tatic</strong></p> <p>Com mais de 15 anos de experiência, a Tatic é uma empresa inovadora de tecnologia da informação. Nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios: uma verdadeira parceria estratégica para o sucesso.</p> <p>Os produtos e serviços Tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados, potencializando os BSS (Business Support Systems) e a aplicação de Machine Learning, Analytics avançado e muito mais para seu negócio.</p> <p>Hoje, grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos, reduzir custos operacionais, reter e fidelizar clientes e aumentar receita. Afinal, a eficiência em dados transforma resultados.</p> <h6><strong>Responsabilidades:</strong></h6> <ul> <li>Se comprometer e melhorar o ciclo de vida dos serviços — desde a concepção e desenho até o deploy e operação;</li> <li>Suportar os serviços antes da entrada em produção;</li> <li>Manter os serviços em operação por meio de métricas e monitoramento relacionados à disponibilidade, latência e saúde geral do sistema;</li> <li>Colaborar com equipes de desenvolvimento para identificar as necessidades técnicas e funcionais dos sistemas;</li> <li>Atender às solicitações e receber feedbacks dos clientes;</li> <li>Contribuir para a melhoria contínua dos sistemas;</li> <li>Elaborar manuais compatíveis com as práticas de SRE no mercado;</li> <li>Monitorar e revisar o sistema de alarmes em operação;</li> <li>Automatizar processos e procedimentos técnicos;</li> <li>Conduzir a resolução de problemas por meio da técnica <em>Blameless postmortems.</em></li> </ul> ## Tatic: <p>Com mais de 15 anos de experiência, a Tatic é uma empresa inovadora de tecnologia da informação. Nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios: uma verdadeira parceria estratégica para o sucesso.</p> <p>Os produtos e serviços Tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados (big data e analytics). E hoje, grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos, reduzir custos operacionais, reter e fidelizar clientes e aumentar receita.</p><a href='https://coodesh.com/empresas/tatic'>Veja mais no site</a> ## Habilidades: - DevOps - Azure - GCP - AWS - Linux - PostgreSQL - Infraestruturas/Arquitetura de redes ## Local: 100% Remoto ## Requisitos: - Cursando ou formado em superior na área de Tecnologia da Informação; - Proatividade, foco e comprometimento com a saúde geral dos sistemas; - Trabalhar em equipe; - Pensamento analítico para prevenção de problemas; - Conhecimentos em sistemas operacionais Linux; - Conhecimento de bash scripts; - Conhecimento de Kubernetes; - Conhecimento em Segurança de Redes, Orquestração de Containers e Serviços de Redes. ## Diferenciais: - Conhecimento de Oracle Linux; - Conhecimento de filesystem (BTRFS e XFS); - Espanhol avançado; - Conhecimento de bases de dados não relacionais; - Experiência com administração de banco de dados PostgreSQL; - Conhecimento de openshift; ## Benefícios: - 100% remoto; - Aulas de Espanhol. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Site Reliability Engineer (SRE) na Tatic](https://coodesh.com/vagas/site-reliability-engineer-sre-180605404?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria DevOps
1.0
[Remoto] Site Reliability Engineer (SRE) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/site-reliability-engineer-sre-180605404?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Tatic</strong> está em busca de <strong><ins>Site Reliability Engineer (SRE)</ins></strong> para compor seu time!</p> <p>Buscamos <strong><em><ins>Pessoas SRE - Site Reliability Engineer</ins></em></strong><em><ins> </ins></em><strong><em><ins>Pleno</ins></em></strong> para atuar 100% remoto na Tatic.&nbsp;</p> <p><strong>Conheça a Tatic</strong></p> <p>Com mais de 15 anos de experiência, a Tatic é uma empresa inovadora de tecnologia da informação. Nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios: uma verdadeira parceria estratégica para o sucesso.</p> <p>Os produtos e serviços Tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados, potencializando os BSS (Business Support Systems) e a aplicação de Machine Learning, Analytics avançado e muito mais para seu negócio.</p> <p>Hoje, grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos, reduzir custos operacionais, reter e fidelizar clientes e aumentar receita. Afinal, a eficiência em dados transforma resultados.</p> <h6><strong>Responsabilidades:</strong></h6> <ul> <li>Se comprometer e melhorar o ciclo de vida dos serviços — desde a concepção e desenho até o deploy e operação;</li> <li>Suportar os serviços antes da entrada em produção;</li> <li>Manter os serviços em operação por meio de métricas e monitoramento relacionados à disponibilidade, latência e saúde geral do sistema;</li> <li>Colaborar com equipes de desenvolvimento para identificar as necessidades técnicas e funcionais dos sistemas;</li> <li>Atender às solicitações e receber feedbacks dos clientes;</li> <li>Contribuir para a melhoria contínua dos sistemas;</li> <li>Elaborar manuais compatíveis com as práticas de SRE no mercado;</li> <li>Monitorar e revisar o sistema de alarmes em operação;</li> <li>Automatizar processos e procedimentos técnicos;</li> <li>Conduzir a resolução de problemas por meio da técnica <em>Blameless postmortems.</em></li> </ul> ## Tatic: <p>Com mais de 15 anos de experiência, a Tatic é uma empresa inovadora de tecnologia da informação. Nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios: uma verdadeira parceria estratégica para o sucesso.</p> <p>Os produtos e serviços Tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados (big data e analytics). E hoje, grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos, reduzir custos operacionais, reter e fidelizar clientes e aumentar receita.</p><a href='https://coodesh.com/empresas/tatic'>Veja mais no site</a> ## Habilidades: - DevOps - Azure - GCP - AWS - Linux - PostgreSQL - Infraestruturas/Arquitetura de redes ## Local: 100% Remoto ## Requisitos: - Cursando ou formado em superior na área de Tecnologia da Informação; - Proatividade, foco e comprometimento com a saúde geral dos sistemas; - Trabalhar em equipe; - Pensamento analítico para prevenção de problemas; - Conhecimentos em sistemas operacionais Linux; - Conhecimento de bash scripts; - Conhecimento de Kubernetes; - Conhecimento em Segurança de Redes, Orquestração de Containers e Serviços de Redes. ## Diferenciais: - Conhecimento de Oracle Linux; - Conhecimento de filesystem (BTRFS e XFS); - Espanhol avançado; - Conhecimento de bases de dados não relacionais; - Experiência com administração de banco de dados PostgreSQL; - Conhecimento de openshift; ## Benefícios: - 100% remoto; - Aulas de Espanhol. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Site Reliability Engineer (SRE) na Tatic](https://coodesh.com/vagas/site-reliability-engineer-sre-180605404?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria DevOps
process
site reliability engineer sre na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a tatic está em busca de site reliability engineer sre para compor seu time buscamos pessoas sre site reliability engineer pleno para atuar remoto na tatic nbsp conheça a tatic com mais de anos de experiência a tatic é uma empresa inovadora de tecnologia da informação nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios uma verdadeira parceria estratégica para o sucesso os produtos e serviços tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados potencializando os bss business support systems e a aplicação de machine learning analytics avançado e muito mais para seu negócio hoje grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos reduzir custos operacionais reter e fidelizar clientes e aumentar receita afinal a eficiência em dados transforma resultados responsabilidades se comprometer e melhorar o ciclo de vida dos serviços — desde a concepção e desenho até o deploy e operação suportar os serviços antes da entrada em produção manter os serviços em operação por meio de métricas e monitoramento relacionados à disponibilidade latência e saúde geral do sistema colaborar com equipes de desenvolvimento para identificar as necessidades técnicas e funcionais dos sistemas atender às solicitações e receber feedbacks dos clientes contribuir para a melhoria contínua dos sistemas elaborar manuais compatíveis com as práticas de sre no mercado monitorar e revisar o sistema de alarmes em operação automatizar processos e procedimentos técnicos conduzir a resolução de problemas por meio da técnica blameless postmortems tatic com mais de anos de experiência a tatic é uma empresa inovadora de tecnologia da informação nossos serviços e soluções irão modernizar sua arquitetura de dados e impulsionar a adoção da inteligência artificial na tomada de decisões de negócios uma verdadeira parceria estratégica para o sucesso os produtos e serviços tatic atendem às necessidades de tratamento e armazenamento de grandes volumes de dados big data e analytics e hoje grandes corporações em nível global já se beneficiam da nossa expertise para otimizar investimentos reduzir custos operacionais reter e fidelizar clientes e aumentar receita habilidades devops azure gcp aws linux postgresql infraestruturas arquitetura de redes local remoto requisitos cursando ou formado em superior na área de tecnologia da informação proatividade foco e comprometimento com a saúde geral dos sistemas trabalhar em equipe pensamento analítico para prevenção de problemas conhecimentos em sistemas operacionais linux conhecimento de bash scripts conhecimento de kubernetes conhecimento em segurança de redes orquestração de containers e serviços de redes diferenciais conhecimento de oracle linux conhecimento de filesystem btrfs e xfs espanhol avançado conhecimento de bases de dados não relacionais experiência com administração de banco de dados postgresql conhecimento de openshift benefícios remoto aulas de espanhol como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime pj categoria devops
1
2,169
5,019,525,145
IssuesEvent
2016-12-14 12:03:39
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Config gone in config manager of abm arakoon
process_wontfix
**On all our CI test environments** we noticed the following exception: ``` [EXCEPTION] Error during execution of <alba_health_check.AlbaHealthCheck object at 0x7fd74ab24510>.run Traceback (most recent call last): File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 35, in <module> class HealthCheckController(object): File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 419, in HealthCheckController HealthCheckController.run_method(*arguments) File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 403, in run_method getattr(cl, option['function'])(logger) File "/opt/OpenvStorage/ovs/extensions/healthcheck/alba/alba_health_check.py", line 675, in run AlbaHealthCheck.get_disk_safety(logger) File "/opt/OpenvStorage/ovs/extensions/healthcheck/alba/alba_health_check.py", line 537, in get_disk_safety namespaces = AlbaCLI.run(command="get-disk-safety", config=config) File "/opt/OpenvStorage/ovs/extensions/healthcheck/helpers/albacli.py", line 156, in run raise RuntimeError(output['error']['message']) RuntimeError: Arakoon_exc.Exception(4, "Not_found") ``` After looking into why the exception occurs I noticed the following: ``` root@ovs-node01-1604:~# alba get-disk-safety --config=arakoon://config/ovs/arakoon/arakoon-mybackend02-abm/config?ini=%2Fopt%2FOpenvStorage%2Fconfig%2Farakoon_cacc.ini 2016-12-14 11:55:08 899347 +0100 - ovs-node01-1604 - 9033/0 - alba/cli - 0 - warning - Arakoon_exc.Exception(4, "Not_found") root@ovs-node01-1604:~# ovs config get /ovs/arakoon/arakoon-mybackend02-abm/config root@ovs-node01-1604:~# ``` Apparently the arakoon ABM config is gone from the config manager.
1.0
Config gone in config manager of abm arakoon - **On all our CI test environments** we noticed the following exception: ``` [EXCEPTION] Error during execution of <alba_health_check.AlbaHealthCheck object at 0x7fd74ab24510>.run Traceback (most recent call last): File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 35, in <module> class HealthCheckController(object): File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 419, in HealthCheckController HealthCheckController.run_method(*arguments) File "/opt/OpenvStorage/ovs/lib/healthcheck.py", line 403, in run_method getattr(cl, option['function'])(logger) File "/opt/OpenvStorage/ovs/extensions/healthcheck/alba/alba_health_check.py", line 675, in run AlbaHealthCheck.get_disk_safety(logger) File "/opt/OpenvStorage/ovs/extensions/healthcheck/alba/alba_health_check.py", line 537, in get_disk_safety namespaces = AlbaCLI.run(command="get-disk-safety", config=config) File "/opt/OpenvStorage/ovs/extensions/healthcheck/helpers/albacli.py", line 156, in run raise RuntimeError(output['error']['message']) RuntimeError: Arakoon_exc.Exception(4, "Not_found") ``` After looking into why the exception occurs I noticed the following: ``` root@ovs-node01-1604:~# alba get-disk-safety --config=arakoon://config/ovs/arakoon/arakoon-mybackend02-abm/config?ini=%2Fopt%2FOpenvStorage%2Fconfig%2Farakoon_cacc.ini 2016-12-14 11:55:08 899347 +0100 - ovs-node01-1604 - 9033/0 - alba/cli - 0 - warning - Arakoon_exc.Exception(4, "Not_found") root@ovs-node01-1604:~# ovs config get /ovs/arakoon/arakoon-mybackend02-abm/config root@ovs-node01-1604:~# ``` Apparently the arakoon ABM config is gone from the config manager.
process
config gone in config manager of abm arakoon on all our ci test environments we noticed the following exception error during execution of run traceback most recent call last file opt openvstorage ovs lib healthcheck py line in class healthcheckcontroller object file opt openvstorage ovs lib healthcheck py line in healthcheckcontroller healthcheckcontroller run method arguments file opt openvstorage ovs lib healthcheck py line in run method getattr cl option logger file opt openvstorage ovs extensions healthcheck alba alba health check py line in run albahealthcheck get disk safety logger file opt openvstorage ovs extensions healthcheck alba alba health check py line in get disk safety namespaces albacli run command get disk safety config config file opt openvstorage ovs extensions healthcheck helpers albacli py line in run raise runtimeerror output runtimeerror arakoon exc exception not found after looking into why the exception occurs i noticed the following root ovs alba get disk safety config arakoon config ovs arakoon arakoon abm config ini cacc ini ovs alba cli warning arakoon exc exception not found root ovs ovs config get ovs arakoon arakoon abm config root ovs apparently the arakoon abm config is gone from the config manager
1
19,914
26,375,608,737
IssuesEvent
2023-01-12 02:00:07
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 12 Jan 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Adapting to Skew: Imputing Spatiotemporal Urban Data with 3D Partial Convolutions and Biased Masking - **Authors:** Bin Han, Bill Howe - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY) - **Arxiv link:** https://arxiv.org/abs/2301.04233 - **Pdf link:** https://arxiv.org/pdf/2301.04233 - **Abstract** We adapt image inpainting techniques to impute large, irregular missing regions in urban settings characterized by sparsity, variance in both space and time, and anomalous events. Missing regions in urban data can be caused by sensor or software failures, data quality issues, interference from weather events, incomplete data collection, or varying data use regulations; any missing data can render the entire dataset unusable for downstream applications. To ensure coverage and utility, we adapt computer vision techniques for image inpainting to operate on 3D histograms (2D space + 1D time) commonly used for data exchange in urban settings. Adapting these techniques to the spatiotemporal setting requires handling skew: urban data tend to follow population density patterns (small dense regions surrounded by large sparse areas); these patterns can dominate the learning process and fool the model into ignoring local or transient effects. To combat skew, we 1) train simultaneously in space and time, and 2) focus attention on dense regions by biasing the masks used for training to the skew in the data. We evaluate the core model and these two extensions using the NYC taxi data and the NYC bikeshare data, simulating different conditions for missing data. We show that the core model is effective qualitatively and quantitatively, and that biased masking during training reduces error in a variety of scenarios. We also articulate a tradeoff in varying the number of timesteps per training sample: too few timesteps and the model ignores transient events; too many timesteps and the model is slow to train with limited performance gain. ### Generic Event Boundary Detection in Video with Pyramid Features - **Authors:** Van Thong Huynh, Hyung-Jeong Yang, Guee-Sang Lee, Soo-Hyung Kim - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04288 - **Pdf link:** https://arxiv.org/pdf/2301.04288 - **Abstract** Generic event boundary detection (GEBD) aims to split video into chunks at a broad and diverse set of actions as humans naturally perceive event boundaries. In this study, we present an approach that considers the correlation between neighbor frames with pyramid feature maps in both spatial and temporal dimensions to construct a framework for localizing generic events in video. The features at multiple spatial dimensions of a pre-trained ResNet-50 are exploited with different views in the temporal dimension to form a temporal pyramid feature map. Based on that, the similarity between neighbor frames is calculated and projected to build a temporal pyramid similarity feature vector. A decoder with 1D convolution operations is used to decode these similarities to a new representation that incorporates their temporal relationship for later boundary score estimation. Extensive experiments conducted on the GEBD benchmark dataset show the effectiveness of our system and its variations, in which we outperformed the state-of-the-art approaches. Additional experiments on TAPOS dataset, which contains long-form videos with Olympic sport actions, demonstrated the effectiveness of our study compared to others. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Deep Learning based Multi-Label Image Classification of Protest Activities - **Authors:** Yingzhou Lu, Kosaku Sato, Jialu Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2301.04212 - **Pdf link:** https://arxiv.org/pdf/2301.04212 - **Abstract** With the rise of internet technology amidst increasing rates of urbanization, sharing information has never been easier thanks to globally-adopted platforms for digital communication. The resulting output of massive amounts of user-generated data can be used to enhance our understanding of significant societal issues particularly for urbanizing areas. In order to better analyze protest behavior, we enhanced the GSR dataset and manually labeled all the images. We used deep learning techniques to analyze social media data to detect social unrest through image classification, which performed good in predict multi-attributes, then also used map visualization to display protest behaviors across the country. ### Elevation Estimation-Driven Building 3D Reconstruction from Single-View Remote Sensing Imagery - **Authors:** Yongqiang Mao, Kaiqiang Chen, Liangjin Zhao, Wei Chen, Deke Tang, Wenjie Liu, Zhirui Wang, Wenhui Diao, Xian Sun, Kun Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04581 - **Pdf link:** https://arxiv.org/pdf/2301.04581 - **Abstract** Building 3D reconstruction from remote sensing images has a wide range of applications in smart cities, photogrammetry and other fields. Methods for automatic 3D urban building modeling typically employ multi-view images as input to algorithms to recover point clouds and 3D models of buildings. However, such models rely heavily on multi-view images of buildings, which are time-intensive and limit the applicability and practicality of the models. To solve these issues, we focus on designing an efficient DSM estimation-driven reconstruction framework (Building3D), which aims to reconstruct 3D building models from the input single-view remote sensing image. First, we propose a Semantic Flow Field-guided DSM Estimation (SFFDE) network, which utilizes the proposed concept of elevation semantic flow to achieve the registration of local and global features. Specifically, in order to make the network semantics globally aware, we propose an Elevation Semantic Globalization (ESG) module to realize the semantic globalization of instances. Further, in order to alleviate the semantic span of global features and original local features, we propose a Local-to-Global Elevation Semantic Registration (L2G-ESR) module based on elevation semantic flow. Our Building3D is rooted in the SFFDE network for building elevation prediction, synchronized with a building extraction network for building masks, and then sequentially performs point cloud reconstruction, surface reconstruction (or CityGML model reconstruction). On this basis, our Building3D can optionally generate CityGML models or surface mesh models of the buildings. Extensive experiments on ISPRS Vaihingen and DFC2019 datasets on the DSM estimation task show that our SFFDE significantly improves upon state-of-the-arts. Furthermore, our Building3D achieves impressive results in the 3D point cloud and 3D model reconstruction process. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Pix2Map: Cross-modal Retrieval for Inferring Street Maps from Images - **Authors:** Xindi Wu, KwunFung Lau, Francesco Ferroni, Aljoša Ošep, Deva Ramanan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2301.04224 - **Pdf link:** https://arxiv.org/pdf/2301.04224 - **Abstract** Self-driving vehicles rely on urban street maps for autonomous navigation. In this paper, we introduce Pix2Map, a method for inferring urban street map topology directly from ego-view images, as needed to continually update and expand existing maps. This is a challenging task, as we need to infer a complex urban road topology directly from raw image data. The main insight of this paper is that this problem can be posed as cross-modal retrieval by learning a joint, cross-modal embedding space for images and existing maps, represented as discrete graphs that encode the topological layout of the visual surroundings. We conduct our experimental evaluation using the Argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data. Moreover, we show that our retrieved maps can be used to update or expand existing maps and even show proof-of-concept results for visual localization and image retrieval from spatial graphs. ### Fast spline detection in high density microscopy data - **Authors:** Albert Alonso, Julius B. Kirkegaard - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Quantitative Methods (q-bio.QM) - **Arxiv link:** https://arxiv.org/abs/2301.04460 - **Pdf link:** https://arxiv.org/pdf/2301.04460 - **Abstract** Computer-aided analysis of biological microscopy data has seen a massive improvement with the utilization of general-purpose deep learning techniques. Yet, in microscopy studies of multi-organism systems, the problem of collision and overlap remains challenging. This is particularly true for systems composed of slender bodies such as crawling nematodes, swimming spermatozoa, or the beating of eukaryotic or prokaryotic flagella. Here, we develop a novel end-to-end deep learning approach to extract precise shape trajectories of generally motile and overlapping splines. Our method works in low resolution settings where feature keypoints are hard to define and detect. Detection is fast and we demonstrate the ability to track thousands of overlapping organisms simultaneously. While our approach is agnostic to area of application, we present it in the setting of and exemplify its usability on dense experiments of crawling Caenorhabditis elegans. The model training is achieved purely on synthetic data, utilizing a physics-based model for nematode motility, and we demonstrate the model's ability to generalize from simulations to experimental videos. ### A new sampling methodology for creating rich, heterogeneous, subsets of samples for training image segmentation algorithms - **Authors:** Matheus Viana da Silva, Natália de Carvalho Santos, Baptiste Lacoste, Cesar Henrique Comin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04517 - **Pdf link:** https://arxiv.org/pdf/2301.04517 - **Abstract** Creating a dataset for training supervised machine learning algorithms can be a demanding task. This is especially true for medical image segmentation since this task usually requires one or more specialists for image annotation, and creating ground truth labels for just a single image can take up to several hours. In addition, it is paramount that the annotated samples represent well the different conditions that might affect the imaged tissue as well as possible changes in the image acquisition process. This can only be achieved by considering samples that are typical in the dataset as well as atypical, or even outlier, samples. We introduce a new sampling methodology for selecting relevant images from a larger non-annotated dataset in a way that evenly considers both prototypical as well as atypical samples. The methodology involves the generation of a uniform grid from a feature space representing the samples, which is then used for randomly drawing relevant images. The selected images provide a uniform cover of the original dataset, and thus define a heterogeneous set of images that can be annotated and used for training supervised segmentation algorithms. We provide a case example by creating a dataset containing a representative set of blood vessel microscopy images selected from a larger dataset containing thousands of images. ## Keyword: raw image ### Pix2Map: Cross-modal Retrieval for Inferring Street Maps from Images - **Authors:** Xindi Wu, KwunFung Lau, Francesco Ferroni, Aljoša Ošep, Deva Ramanan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2301.04224 - **Pdf link:** https://arxiv.org/pdf/2301.04224 - **Abstract** Self-driving vehicles rely on urban street maps for autonomous navigation. In this paper, we introduce Pix2Map, a method for inferring urban street map topology directly from ego-view images, as needed to continually update and expand existing maps. This is a challenging task, as we need to infer a complex urban road topology directly from raw image data. The main insight of this paper is that this problem can be posed as cross-modal retrieval by learning a joint, cross-modal embedding space for images and existing maps, represented as discrete graphs that encode the topological layout of the visual surroundings. We conduct our experimental evaluation using the Argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data. Moreover, we show that our retrieved maps can be used to update or expand existing maps and even show proof-of-concept results for visual localization and image retrieval from spatial graphs.
2.0
New submissions for Thu, 12 Jan 23 - ## Keyword: events ### Adapting to Skew: Imputing Spatiotemporal Urban Data with 3D Partial Convolutions and Biased Masking - **Authors:** Bin Han, Bill Howe - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY) - **Arxiv link:** https://arxiv.org/abs/2301.04233 - **Pdf link:** https://arxiv.org/pdf/2301.04233 - **Abstract** We adapt image inpainting techniques to impute large, irregular missing regions in urban settings characterized by sparsity, variance in both space and time, and anomalous events. Missing regions in urban data can be caused by sensor or software failures, data quality issues, interference from weather events, incomplete data collection, or varying data use regulations; any missing data can render the entire dataset unusable for downstream applications. To ensure coverage and utility, we adapt computer vision techniques for image inpainting to operate on 3D histograms (2D space + 1D time) commonly used for data exchange in urban settings. Adapting these techniques to the spatiotemporal setting requires handling skew: urban data tend to follow population density patterns (small dense regions surrounded by large sparse areas); these patterns can dominate the learning process and fool the model into ignoring local or transient effects. To combat skew, we 1) train simultaneously in space and time, and 2) focus attention on dense regions by biasing the masks used for training to the skew in the data. We evaluate the core model and these two extensions using the NYC taxi data and the NYC bikeshare data, simulating different conditions for missing data. We show that the core model is effective qualitatively and quantitatively, and that biased masking during training reduces error in a variety of scenarios. We also articulate a tradeoff in varying the number of timesteps per training sample: too few timesteps and the model ignores transient events; too many timesteps and the model is slow to train with limited performance gain. ### Generic Event Boundary Detection in Video with Pyramid Features - **Authors:** Van Thong Huynh, Hyung-Jeong Yang, Guee-Sang Lee, Soo-Hyung Kim - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04288 - **Pdf link:** https://arxiv.org/pdf/2301.04288 - **Abstract** Generic event boundary detection (GEBD) aims to split video into chunks at a broad and diverse set of actions as humans naturally perceive event boundaries. In this study, we present an approach that considers the correlation between neighbor frames with pyramid feature maps in both spatial and temporal dimensions to construct a framework for localizing generic events in video. The features at multiple spatial dimensions of a pre-trained ResNet-50 are exploited with different views in the temporal dimension to form a temporal pyramid feature map. Based on that, the similarity between neighbor frames is calculated and projected to build a temporal pyramid similarity feature vector. A decoder with 1D convolution operations is used to decode these similarities to a new representation that incorporates their temporal relationship for later boundary score estimation. Extensive experiments conducted on the GEBD benchmark dataset show the effectiveness of our system and its variations, in which we outperformed the state-of-the-art approaches. Additional experiments on TAPOS dataset, which contains long-form videos with Olympic sport actions, demonstrated the effectiveness of our study compared to others. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Deep Learning based Multi-Label Image Classification of Protest Activities - **Authors:** Yingzhou Lu, Kosaku Sato, Jialu Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2301.04212 - **Pdf link:** https://arxiv.org/pdf/2301.04212 - **Abstract** With the rise of internet technology amidst increasing rates of urbanization, sharing information has never been easier thanks to globally-adopted platforms for digital communication. The resulting output of massive amounts of user-generated data can be used to enhance our understanding of significant societal issues particularly for urbanizing areas. In order to better analyze protest behavior, we enhanced the GSR dataset and manually labeled all the images. We used deep learning techniques to analyze social media data to detect social unrest through image classification, which performed good in predict multi-attributes, then also used map visualization to display protest behaviors across the country. ### Elevation Estimation-Driven Building 3D Reconstruction from Single-View Remote Sensing Imagery - **Authors:** Yongqiang Mao, Kaiqiang Chen, Liangjin Zhao, Wei Chen, Deke Tang, Wenjie Liu, Zhirui Wang, Wenhui Diao, Xian Sun, Kun Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04581 - **Pdf link:** https://arxiv.org/pdf/2301.04581 - **Abstract** Building 3D reconstruction from remote sensing images has a wide range of applications in smart cities, photogrammetry and other fields. Methods for automatic 3D urban building modeling typically employ multi-view images as input to algorithms to recover point clouds and 3D models of buildings. However, such models rely heavily on multi-view images of buildings, which are time-intensive and limit the applicability and practicality of the models. To solve these issues, we focus on designing an efficient DSM estimation-driven reconstruction framework (Building3D), which aims to reconstruct 3D building models from the input single-view remote sensing image. First, we propose a Semantic Flow Field-guided DSM Estimation (SFFDE) network, which utilizes the proposed concept of elevation semantic flow to achieve the registration of local and global features. Specifically, in order to make the network semantics globally aware, we propose an Elevation Semantic Globalization (ESG) module to realize the semantic globalization of instances. Further, in order to alleviate the semantic span of global features and original local features, we propose a Local-to-Global Elevation Semantic Registration (L2G-ESR) module based on elevation semantic flow. Our Building3D is rooted in the SFFDE network for building elevation prediction, synchronized with a building extraction network for building masks, and then sequentially performs point cloud reconstruction, surface reconstruction (or CityGML model reconstruction). On this basis, our Building3D can optionally generate CityGML models or surface mesh models of the buildings. Extensive experiments on ISPRS Vaihingen and DFC2019 datasets on the DSM estimation task show that our SFFDE significantly improves upon state-of-the-arts. Furthermore, our Building3D achieves impressive results in the 3D point cloud and 3D model reconstruction process. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Pix2Map: Cross-modal Retrieval for Inferring Street Maps from Images - **Authors:** Xindi Wu, KwunFung Lau, Francesco Ferroni, Aljoša Ošep, Deva Ramanan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2301.04224 - **Pdf link:** https://arxiv.org/pdf/2301.04224 - **Abstract** Self-driving vehicles rely on urban street maps for autonomous navigation. In this paper, we introduce Pix2Map, a method for inferring urban street map topology directly from ego-view images, as needed to continually update and expand existing maps. This is a challenging task, as we need to infer a complex urban road topology directly from raw image data. The main insight of this paper is that this problem can be posed as cross-modal retrieval by learning a joint, cross-modal embedding space for images and existing maps, represented as discrete graphs that encode the topological layout of the visual surroundings. We conduct our experimental evaluation using the Argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data. Moreover, we show that our retrieved maps can be used to update or expand existing maps and even show proof-of-concept results for visual localization and image retrieval from spatial graphs. ### Fast spline detection in high density microscopy data - **Authors:** Albert Alonso, Julius B. Kirkegaard - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Quantitative Methods (q-bio.QM) - **Arxiv link:** https://arxiv.org/abs/2301.04460 - **Pdf link:** https://arxiv.org/pdf/2301.04460 - **Abstract** Computer-aided analysis of biological microscopy data has seen a massive improvement with the utilization of general-purpose deep learning techniques. Yet, in microscopy studies of multi-organism systems, the problem of collision and overlap remains challenging. This is particularly true for systems composed of slender bodies such as crawling nematodes, swimming spermatozoa, or the beating of eukaryotic or prokaryotic flagella. Here, we develop a novel end-to-end deep learning approach to extract precise shape trajectories of generally motile and overlapping splines. Our method works in low resolution settings where feature keypoints are hard to define and detect. Detection is fast and we demonstrate the ability to track thousands of overlapping organisms simultaneously. While our approach is agnostic to area of application, we present it in the setting of and exemplify its usability on dense experiments of crawling Caenorhabditis elegans. The model training is achieved purely on synthetic data, utilizing a physics-based model for nematode motility, and we demonstrate the model's ability to generalize from simulations to experimental videos. ### A new sampling methodology for creating rich, heterogeneous, subsets of samples for training image segmentation algorithms - **Authors:** Matheus Viana da Silva, Natália de Carvalho Santos, Baptiste Lacoste, Cesar Henrique Comin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.04517 - **Pdf link:** https://arxiv.org/pdf/2301.04517 - **Abstract** Creating a dataset for training supervised machine learning algorithms can be a demanding task. This is especially true for medical image segmentation since this task usually requires one or more specialists for image annotation, and creating ground truth labels for just a single image can take up to several hours. In addition, it is paramount that the annotated samples represent well the different conditions that might affect the imaged tissue as well as possible changes in the image acquisition process. This can only be achieved by considering samples that are typical in the dataset as well as atypical, or even outlier, samples. We introduce a new sampling methodology for selecting relevant images from a larger non-annotated dataset in a way that evenly considers both prototypical as well as atypical samples. The methodology involves the generation of a uniform grid from a feature space representing the samples, which is then used for randomly drawing relevant images. The selected images provide a uniform cover of the original dataset, and thus define a heterogeneous set of images that can be annotated and used for training supervised segmentation algorithms. We provide a case example by creating a dataset containing a representative set of blood vessel microscopy images selected from a larger dataset containing thousands of images. ## Keyword: raw image ### Pix2Map: Cross-modal Retrieval for Inferring Street Maps from Images - **Authors:** Xindi Wu, KwunFung Lau, Francesco Ferroni, Aljoša Ošep, Deva Ramanan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2301.04224 - **Pdf link:** https://arxiv.org/pdf/2301.04224 - **Abstract** Self-driving vehicles rely on urban street maps for autonomous navigation. In this paper, we introduce Pix2Map, a method for inferring urban street map topology directly from ego-view images, as needed to continually update and expand existing maps. This is a challenging task, as we need to infer a complex urban road topology directly from raw image data. The main insight of this paper is that this problem can be posed as cross-modal retrieval by learning a joint, cross-modal embedding space for images and existing maps, represented as discrete graphs that encode the topological layout of the visual surroundings. We conduct our experimental evaluation using the Argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data. Moreover, we show that our retrieved maps can be used to update or expand existing maps and even show proof-of-concept results for visual localization and image retrieval from spatial graphs.
process
new submissions for thu jan keyword events adapting to skew imputing spatiotemporal urban data with partial convolutions and biased masking authors bin han bill howe subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computers and society cs cy arxiv link pdf link abstract we adapt image inpainting techniques to impute large irregular missing regions in urban settings characterized by sparsity variance in both space and time and anomalous events missing regions in urban data can be caused by sensor or software failures data quality issues interference from weather events incomplete data collection or varying data use regulations any missing data can render the entire dataset unusable for downstream applications to ensure coverage and utility we adapt computer vision techniques for image inpainting to operate on histograms space time commonly used for data exchange in urban settings adapting these techniques to the spatiotemporal setting requires handling skew urban data tend to follow population density patterns small dense regions surrounded by large sparse areas these patterns can dominate the learning process and fool the model into ignoring local or transient effects to combat skew we train simultaneously in space and time and focus attention on dense regions by biasing the masks used for training to the skew in the data we evaluate the core model and these two extensions using the nyc taxi data and the nyc bikeshare data simulating different conditions for missing data we show that the core model is effective qualitatively and quantitatively and that biased masking during training reduces error in a variety of scenarios we also articulate a tradeoff in varying the number of timesteps per training sample too few timesteps and the model ignores transient events too many timesteps and the model is slow to train with limited performance gain generic event boundary detection in video with pyramid features authors van thong huynh hyung jeong yang guee sang lee soo hyung kim subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract generic event boundary detection gebd aims to split video into chunks at a broad and diverse set of actions as humans naturally perceive event boundaries in this study we present an approach that considers the correlation between neighbor frames with pyramid feature maps in both spatial and temporal dimensions to construct a framework for localizing generic events in video the features at multiple spatial dimensions of a pre trained resnet are exploited with different views in the temporal dimension to form a temporal pyramid feature map based on that the similarity between neighbor frames is calculated and projected to build a temporal pyramid similarity feature vector a decoder with convolution operations is used to decode these similarities to a new representation that incorporates their temporal relationship for later boundary score estimation extensive experiments conducted on the gebd benchmark dataset show the effectiveness of our system and its variations in which we outperformed the state of the art approaches additional experiments on tapos dataset which contains long form videos with olympic sport actions demonstrated the effectiveness of our study compared to others keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp deep learning based multi label image classification of protest activities authors yingzhou lu kosaku sato jialu wang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract with the rise of internet technology amidst increasing rates of urbanization sharing information has never been easier thanks to globally adopted platforms for digital communication the resulting output of massive amounts of user generated data can be used to enhance our understanding of significant societal issues particularly for urbanizing areas in order to better analyze protest behavior we enhanced the gsr dataset and manually labeled all the images we used deep learning techniques to analyze social media data to detect social unrest through image classification which performed good in predict multi attributes then also used map visualization to display protest behaviors across the country elevation estimation driven building reconstruction from single view remote sensing imagery authors yongqiang mao kaiqiang chen liangjin zhao wei chen deke tang wenjie liu zhirui wang wenhui diao xian sun kun fu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract building reconstruction from remote sensing images has a wide range of applications in smart cities photogrammetry and other fields methods for automatic urban building modeling typically employ multi view images as input to algorithms to recover point clouds and models of buildings however such models rely heavily on multi view images of buildings which are time intensive and limit the applicability and practicality of the models to solve these issues we focus on designing an efficient dsm estimation driven reconstruction framework which aims to reconstruct building models from the input single view remote sensing image first we propose a semantic flow field guided dsm estimation sffde network which utilizes the proposed concept of elevation semantic flow to achieve the registration of local and global features specifically in order to make the network semantics globally aware we propose an elevation semantic globalization esg module to realize the semantic globalization of instances further in order to alleviate the semantic span of global features and original local features we propose a local to global elevation semantic registration esr module based on elevation semantic flow our is rooted in the sffde network for building elevation prediction synchronized with a building extraction network for building masks and then sequentially performs point cloud reconstruction surface reconstruction or citygml model reconstruction on this basis our can optionally generate citygml models or surface mesh models of the buildings extensive experiments on isprs vaihingen and datasets on the dsm estimation task show that our sffde significantly improves upon state of the arts furthermore our achieves impressive results in the point cloud and model reconstruction process keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw cross modal retrieval for inferring street maps from images authors xindi wu kwunfung lau francesco ferroni aljoša ošep deva ramanan subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract self driving vehicles rely on urban street maps for autonomous navigation in this paper we introduce a method for inferring urban street map topology directly from ego view images as needed to continually update and expand existing maps this is a challenging task as we need to infer a complex urban road topology directly from raw image data the main insight of this paper is that this problem can be posed as cross modal retrieval by learning a joint cross modal embedding space for images and existing maps represented as discrete graphs that encode the topological layout of the visual surroundings we conduct our experimental evaluation using the argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data moreover we show that our retrieved maps can be used to update or expand existing maps and even show proof of concept results for visual localization and image retrieval from spatial graphs fast spline detection in high density microscopy data authors albert alonso julius b kirkegaard subjects computer vision and pattern recognition cs cv quantitative methods q bio qm arxiv link pdf link abstract computer aided analysis of biological microscopy data has seen a massive improvement with the utilization of general purpose deep learning techniques yet in microscopy studies of multi organism systems the problem of collision and overlap remains challenging this is particularly true for systems composed of slender bodies such as crawling nematodes swimming spermatozoa or the beating of eukaryotic or prokaryotic flagella here we develop a novel end to end deep learning approach to extract precise shape trajectories of generally motile and overlapping splines our method works in low resolution settings where feature keypoints are hard to define and detect detection is fast and we demonstrate the ability to track thousands of overlapping organisms simultaneously while our approach is agnostic to area of application we present it in the setting of and exemplify its usability on dense experiments of crawling caenorhabditis elegans the model training is achieved purely on synthetic data utilizing a physics based model for nematode motility and we demonstrate the model s ability to generalize from simulations to experimental videos a new sampling methodology for creating rich heterogeneous subsets of samples for training image segmentation algorithms authors matheus viana da silva natália de carvalho santos baptiste lacoste cesar henrique comin subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract creating a dataset for training supervised machine learning algorithms can be a demanding task this is especially true for medical image segmentation since this task usually requires one or more specialists for image annotation and creating ground truth labels for just a single image can take up to several hours in addition it is paramount that the annotated samples represent well the different conditions that might affect the imaged tissue as well as possible changes in the image acquisition process this can only be achieved by considering samples that are typical in the dataset as well as atypical or even outlier samples we introduce a new sampling methodology for selecting relevant images from a larger non annotated dataset in a way that evenly considers both prototypical as well as atypical samples the methodology involves the generation of a uniform grid from a feature space representing the samples which is then used for randomly drawing relevant images the selected images provide a uniform cover of the original dataset and thus define a heterogeneous set of images that can be annotated and used for training supervised segmentation algorithms we provide a case example by creating a dataset containing a representative set of blood vessel microscopy images selected from a larger dataset containing thousands of images keyword raw image cross modal retrieval for inferring street maps from images authors xindi wu kwunfung lau francesco ferroni aljoša ošep deva ramanan subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract self driving vehicles rely on urban street maps for autonomous navigation in this paper we introduce a method for inferring urban street map topology directly from ego view images as needed to continually update and expand existing maps this is a challenging task as we need to infer a complex urban road topology directly from raw image data the main insight of this paper is that this problem can be posed as cross modal retrieval by learning a joint cross modal embedding space for images and existing maps represented as discrete graphs that encode the topological layout of the visual surroundings we conduct our experimental evaluation using the argoverse dataset and show that it is indeed possible to accurately retrieve street maps corresponding to both seen and unseen roads solely from image data moreover we show that our retrieved maps can be used to update or expand existing maps and even show proof of concept results for visual localization and image retrieval from spatial graphs
1
21,705
3,548,146,128
IssuesEvent
2016-01-20 13:14:15
briandonahue/FluxJpeg.Core
https://api.github.com/repos/briandonahue/FluxJpeg.Core
opened
Encoding a 8-bit Grayscale jpg fails (colorspace conversion bug) | Resubmit
auto-migrated Priority-High Type-Defect
_From @GoogleCodeExporter on January 20, 2016 12:59_ ``` What steps will reproduce the problem? 1. Try to encode a 8-bit Grayscale jpg What is the expected output? What do you see instead? Expected: Encoded image. Actual: IndexOutOfException on Encoding (Line 261, "CompressTo"-method). What version of the product are you using? On what operating system? Original version (v1.0?) on. Vista Business SP1 Please provide any additional information below. This is a Resubmit with more info. I first reported it as a comment on issue 1: --- Comment 1 by snakeware95, Feb 03, 2009 Also there is a bug in the JpegEncoder constructor converting (8-bit) Grayscale images. Replace: _input.Image.ChangeColorSpace(ColorSpace.YCbCr); With this to fix: _input = new DecodedJpeg(_input.Image.ChangeColorSpace(ColorSpace.YCbCr), _input.MetaHeaders); --- Comment 3 by jeff.powers, Feb 19, 2009 snakeware95: the colorspace change happens in-place so I think that edit isn't needed. resubmit a bug if problems remain. --- The colorspace change itself is not the problem, but the properties of the surrounding DecodedJpeg (BlockWidth array f.e.) are not updated (still 1 component). So when you loop over them using the Image.ComponentCount (which is 3 after the conversion) in "CompressTo", you get an IndexOutOfException on Encoding (Line 261, "CompressTo"-method). Please try it yourself! ``` Original issue reported on code.google.com by `snakewar...@gmail.com` on 3 Mar 2009 at 12:48 _Copied from original issue: anders9ustafsson/fjcore#5_
1.0
Encoding a 8-bit Grayscale jpg fails (colorspace conversion bug) | Resubmit - _From @GoogleCodeExporter on January 20, 2016 12:59_ ``` What steps will reproduce the problem? 1. Try to encode a 8-bit Grayscale jpg What is the expected output? What do you see instead? Expected: Encoded image. Actual: IndexOutOfException on Encoding (Line 261, "CompressTo"-method). What version of the product are you using? On what operating system? Original version (v1.0?) on. Vista Business SP1 Please provide any additional information below. This is a Resubmit with more info. I first reported it as a comment on issue 1: --- Comment 1 by snakeware95, Feb 03, 2009 Also there is a bug in the JpegEncoder constructor converting (8-bit) Grayscale images. Replace: _input.Image.ChangeColorSpace(ColorSpace.YCbCr); With this to fix: _input = new DecodedJpeg(_input.Image.ChangeColorSpace(ColorSpace.YCbCr), _input.MetaHeaders); --- Comment 3 by jeff.powers, Feb 19, 2009 snakeware95: the colorspace change happens in-place so I think that edit isn't needed. resubmit a bug if problems remain. --- The colorspace change itself is not the problem, but the properties of the surrounding DecodedJpeg (BlockWidth array f.e.) are not updated (still 1 component). So when you loop over them using the Image.ComponentCount (which is 3 after the conversion) in "CompressTo", you get an IndexOutOfException on Encoding (Line 261, "CompressTo"-method). Please try it yourself! ``` Original issue reported on code.google.com by `snakewar...@gmail.com` on 3 Mar 2009 at 12:48 _Copied from original issue: anders9ustafsson/fjcore#5_
non_process
encoding a bit grayscale jpg fails colorspace conversion bug resubmit from googlecodeexporter on january what steps will reproduce the problem try to encode a bit grayscale jpg what is the expected output what do you see instead expected encoded image actual indexoutofexception on encoding line compressto method what version of the product are you using on what operating system original version on vista business please provide any additional information below this is a resubmit with more info i first reported it as a comment on issue comment by feb also there is a bug in the jpegencoder constructor converting bit grayscale images replace input image changecolorspace colorspace ycbcr with this to fix input new decodedjpeg input image changecolorspace colorspace ycbcr input metaheaders comment by jeff powers feb the colorspace change happens in place so i think that edit isn t needed resubmit a bug if problems remain the colorspace change itself is not the problem but the properties of the surrounding decodedjpeg blockwidth array f e are not updated still component so when you loop over them using the image componentcount which is after the conversion in compressto you get an indexoutofexception on encoding line compressto method please try it yourself original issue reported on code google com by snakewar gmail com on mar at copied from original issue fjcore
0
585,791
17,534,282,915
IssuesEvent
2021-08-12 03:37:55
shoepro/server
https://api.github.com/repos/shoepro/server
opened
[Feat]: Create user model CRUD logic
Priority: Middle 2.0h Feature Server
### ISSUE - Group: `Server` - Type: `Feat` - Time: `2.0h` - Priority: `Middle` ### TODO 1. [ ] Create user repository file 2. [ ] Create CRUD logic based on API docs
1.0
[Feat]: Create user model CRUD logic - ### ISSUE - Group: `Server` - Type: `Feat` - Time: `2.0h` - Priority: `Middle` ### TODO 1. [ ] Create user repository file 2. [ ] Create CRUD logic based on API docs
non_process
create user model crud logic issue group server type feat time priority middle todo create user repository file create crud logic based on api docs
0
209,623
23,730,711,153
IssuesEvent
2022-08-31 01:16:09
benlazarine/atmosphere-v2-apiary-docs
https://api.github.com/repos/benlazarine/atmosphere-v2-apiary-docs
opened
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benlazarine/atmosphere-v2-apiary-docs/commit/725a188d8ec441be072e185a12d8d7862b8ebd57">725a188d8ec441be072e185a12d8d7862b8ebd57</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benlazarine/atmosphere-v2-apiary-docs/commit/725a188d8ec441be072e185a12d8d7862b8ebd57">725a188d8ec441be072e185a12d8d7862b8ebd57</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node modules redeyed examples browser index html path to vulnerable library node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend
0
3,231
6,289,280,093
IssuesEvent
2017-07-19 18:51:20
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
[S.D.Process] CloseMainWindow_NotStarted_ThrowsInvalidOperationException does not throw
area-System.Diagnostics.Process
(Test case will be added soon, creating issue so that I can disable that in the PR) I believe we should be able to throw even on UAP since we can detect that process is not started. On the other hand we might need to be consistent with non throwing behavior because MainWindowHandle is not supported ``` ERROR: System.Diagnostics.Tests.ProcessTests.CloseMainWindow_NotStarted_ThrowsInvalidOperationException [FAIL] Assert.Throws() Failure Expected: typeof(System.InvalidOperationException) Actual: (No exception was thrown) ```
1.0
[S.D.Process] CloseMainWindow_NotStarted_ThrowsInvalidOperationException does not throw - (Test case will be added soon, creating issue so that I can disable that in the PR) I believe we should be able to throw even on UAP since we can detect that process is not started. On the other hand we might need to be consistent with non throwing behavior because MainWindowHandle is not supported ``` ERROR: System.Diagnostics.Tests.ProcessTests.CloseMainWindow_NotStarted_ThrowsInvalidOperationException [FAIL] Assert.Throws() Failure Expected: typeof(System.InvalidOperationException) Actual: (No exception was thrown) ```
process
closemainwindow notstarted throwsinvalidoperationexception does not throw test case will be added soon creating issue so that i can disable that in the pr i believe we should be able to throw even on uap since we can detect that process is not started on the other hand we might need to be consistent with non throwing behavior because mainwindowhandle is not supported error system diagnostics tests processtests closemainwindow notstarted throwsinvalidoperationexception assert throws failure expected typeof system invalidoperationexception actual no exception was thrown
1
19,586
25,922,149,238
IssuesEvent
2022-12-15 23:17:38
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
closed
Processing > Rebin should work on 1D data
type - enhancement level - easy f - processing
Unfortunately this will most likely require a separate menu item for now; the ability of a computation to make its UI depend on the rank of the data is not possible at the moment. See also - #274
1.0
Processing > Rebin should work on 1D data - Unfortunately this will most likely require a separate menu item for now; the ability of a computation to make its UI depend on the rank of the data is not possible at the moment. See also - #274
process
processing rebin should work on data unfortunately this will most likely require a separate menu item for now the ability of a computation to make its ui depend on the rank of the data is not possible at the moment see also
1
28,506
5,510,585,062
IssuesEvent
2017-03-17 00:30:27
joshcummingsdesign/grizzly-wp
https://api.github.com/repos/joshcummingsdesign/grizzly-wp
closed
Match Cloudways Server
documentation enhancement pipeline
Local and dev should match Cloudways as closely as possible (PHP and Apache configuration, database config, etc.)
1.0
Match Cloudways Server - Local and dev should match Cloudways as closely as possible (PHP and Apache configuration, database config, etc.)
non_process
match cloudways server local and dev should match cloudways as closely as possible php and apache configuration database config etc
0
19,745
26,107,931,275
IssuesEvent
2022-12-27 15:32:39
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/logstransformprocessor] Processor hangs waiting for logs that were filtered out
bug priority:p2 processor/logstransform
### What happened? ## Description When using the `logstransformprocessor` to filter logs, the [loop](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/logstransformprocessor/processor.go#L145) in `processLogs` will eventually hang waiting for the filtered out logs. `processLogs` is blocking, so the pipeline is blocked there. ## Steps to Reproduce Using the supplied config: 1. Use some tool, e.g. [flog](https://github.com/mingrammer/flog) to write logs to `/tmp/test.log` at a constant rate. ```sh flog -f json -o /tmp/test.log -t log -w -l --delay=500ms ``` 2. Run the collector, and observe the exporter stop sending logs after several seconds. ## Expected Result Logs would continue to flow despite some records being dropped by the processor. ## Actual Result All pipelines using the `filelog` receiver stall once the `processLogs` method gets stuck waiting on the channel. The exporter doesn't send any logs despite new records being written constantly. ### Collector version c62e003569096531872cab77c20991e1a5135560 ### Environment information ## Environment OS: Ubuntu 20.04, macOS 12.6 Compiler(if manually compiled): go 1.18.7 ### OpenTelemetry Collector configuration ```yaml receivers: filelog: include: - /tmp/test.log processors: logstransform: operators: - drop_ratio: 0.5 expr: "true" type: filter exporters: otlphttp: endpoint: http://localhost:9123 service: pipelines: logs: receivers: [filelog] processors: [logstransform] exporters: [otlphttp] ``` ### Log output ```shell Processor forwarding one log, then getting stalled: 2022-10-21T09:36:37.178-0400 info pipelines/pipelines.go:106 Receiver started. {"kind": "receiver", "name": "filelog", "pipeline": "logs"} 2022-10-21T09:36:37.178-0400 info service/service.go:105 Everything is ready. Begin running and processing data. 2022-10-21T09:36:37.379-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:37.379-0400 info fileconsumer/file.go:161 Started watching file from end. To read preexisting logs, configure the argument 'start_at' to 'beginning' {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer", "path": "/tmp/test.log"} 2022-10-21T09:36:37.779-0400 debug otlphttpexporter/otlp.go:127 Preparing to make HTTP request {"kind": "exporter", "data_type": "logs", "name": "otlphttp", "url": "http://localhost:9123/v1/logs"} 2022-10-21T09:36:38.380-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:39.579-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:40.579-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:41.779-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:42.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:43.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:44.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} ^C2022-10-21T09:36:45.208-0400 info service/collector.go:193 Received signal from OS {"signal": "interrupt"} 2022-10-21T09:36:45.208-0400 info service/service.go:114 Starting shutdown... ``` ``` ### Additional context _No response_
1.0
[processor/logstransformprocessor] Processor hangs waiting for logs that were filtered out - ### What happened? ## Description When using the `logstransformprocessor` to filter logs, the [loop](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/logstransformprocessor/processor.go#L145) in `processLogs` will eventually hang waiting for the filtered out logs. `processLogs` is blocking, so the pipeline is blocked there. ## Steps to Reproduce Using the supplied config: 1. Use some tool, e.g. [flog](https://github.com/mingrammer/flog) to write logs to `/tmp/test.log` at a constant rate. ```sh flog -f json -o /tmp/test.log -t log -w -l --delay=500ms ``` 2. Run the collector, and observe the exporter stop sending logs after several seconds. ## Expected Result Logs would continue to flow despite some records being dropped by the processor. ## Actual Result All pipelines using the `filelog` receiver stall once the `processLogs` method gets stuck waiting on the channel. The exporter doesn't send any logs despite new records being written constantly. ### Collector version c62e003569096531872cab77c20991e1a5135560 ### Environment information ## Environment OS: Ubuntu 20.04, macOS 12.6 Compiler(if manually compiled): go 1.18.7 ### OpenTelemetry Collector configuration ```yaml receivers: filelog: include: - /tmp/test.log processors: logstransform: operators: - drop_ratio: 0.5 expr: "true" type: filter exporters: otlphttp: endpoint: http://localhost:9123 service: pipelines: logs: receivers: [filelog] processors: [logstransform] exporters: [otlphttp] ``` ### Log output ```shell Processor forwarding one log, then getting stalled: 2022-10-21T09:36:37.178-0400 info pipelines/pipelines.go:106 Receiver started. {"kind": "receiver", "name": "filelog", "pipeline": "logs"} 2022-10-21T09:36:37.178-0400 info service/service.go:105 Everything is ready. Begin running and processing data. 2022-10-21T09:36:37.379-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:37.379-0400 info fileconsumer/file.go:161 Started watching file from end. To read preexisting logs, configure the argument 'start_at' to 'beginning' {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer", "path": "/tmp/test.log"} 2022-10-21T09:36:37.779-0400 debug otlphttpexporter/otlp.go:127 Preparing to make HTTP request {"kind": "exporter", "data_type": "logs", "name": "otlphttp", "url": "http://localhost:9123/v1/logs"} 2022-10-21T09:36:38.380-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:39.579-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:40.579-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:41.779-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:42.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:43.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} 2022-10-21T09:36:44.979-0400 debug fileconsumer/file.go:124 Consuming files {"kind": "receiver", "name": "filelog", "pipeline": "logs", "component": "fileconsumer"} ^C2022-10-21T09:36:45.208-0400 info service/collector.go:193 Received signal from OS {"signal": "interrupt"} 2022-10-21T09:36:45.208-0400 info service/service.go:114 Starting shutdown... ``` ``` ### Additional context _No response_
process
processor hangs waiting for logs that were filtered out what happened description when using the logstransformprocessor to filter logs the in processlogs will eventually hang waiting for the filtered out logs processlogs is blocking so the pipeline is blocked there steps to reproduce using the supplied config use some tool e g to write logs to tmp test log at a constant rate sh flog f json o tmp test log t log w l delay run the collector and observe the exporter stop sending logs after several seconds expected result logs would continue to flow despite some records being dropped by the processor actual result all pipelines using the filelog receiver stall once the processlogs method gets stuck waiting on the channel the exporter doesn t send any logs despite new records being written constantly collector version environment information environment os ubuntu macos compiler if manually compiled go opentelemetry collector configuration yaml receivers filelog include tmp test log processors logstransform operators drop ratio expr true type filter exporters otlphttp endpoint service pipelines logs receivers processors exporters log output shell processor forwarding one log then getting stalled info pipelines pipelines go receiver started kind receiver name filelog pipeline logs info service service go everything is ready begin running and processing data debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer info fileconsumer file go started watching file from end to read preexisting logs configure the argument start at to beginning kind receiver name filelog pipeline logs component fileconsumer path tmp test log debug otlphttpexporter otlp go preparing to make http request kind exporter data type logs name otlphttp url debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer debug fileconsumer file go consuming files kind receiver name filelog pipeline logs component fileconsumer info service collector go received signal from os signal interrupt info service service go starting shutdown additional context no response
1
49,079
13,440,480,427
IssuesEvent
2020-09-08 01:03:27
jgeraigery/crnk-framework
https://api.github.com/repos/jgeraigery/crnk-framework
opened
CVE-2020-24616 (High) detected in jackson-databind-2.9.10.jar
security vulnerability
## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar</p> <p> Dependency Hierarchy: - json-schema-validator-2.9.0.jar (Root Library) - json-schema-validator-2.2.6.jar - json-schema-core-1.2.5.jar - jackson-coreutils-1.8.jar - :x: **jackson-databind-2.9.10.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10","isTransitiveDependency":true,"dependencyTree":"com.jayway.restassured:json-schema-validator:2.9.0;com.github.fge:json-schema-validator:2.2.6;com.github.fge:json-schema-core:1.2.5;com.github.fge:jackson-coreutils:1.8;com.fasterxml.jackson.core:jackson-databind:2.9.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-24616 (High) detected in jackson-databind-2.9.10.jar - ## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar</p> <p> Dependency Hierarchy: - json-schema-validator-2.9.0.jar (Root Library) - json-schema-validator-2.2.6.jar - json-schema-core-1.2.5.jar - jackson-coreutils-1.8.jar - :x: **jackson-databind-2.9.10.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10","isTransitiveDependency":true,"dependencyTree":"com.jayway.restassured:json-schema-validator:2.9.0;com.github.fge:json-schema-validator:2.2.6;com.github.fge:json-schema-core:1.2.5;com.github.fge:jackson-coreutils:1.8;com.fasterxml.jackson.core:jackson-databind:2.9.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar le caches modules files com fasterxml jackson core jackson databind jackson databind jar le caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy json schema validator jar root library json schema validator jar json schema core jar jackson coreutils jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp vulnerabilityurl
0
328,358
28,116,372,598
IssuesEvent
2023-03-31 11:04:16
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
sql: TestQueryCache failed
C-test-failure O-robot branch-release-23.1
sql.TestQueryCache [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9357134?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9357134?buildTab=artifacts#/) on release-23.1 @ [7e72aae900c3ff4b44f1643c2d7ba55fbb2cbe23](https://github.com/cockroachdb/cockroach/commits/7e72aae900c3ff4b44f1643c2d7ba55fbb2cbe23): ``` goroutine 204895147 lock 0xc0037d79d0 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:1186 kvserver.(*Replica).tick ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:1185 kvserver.(*Replica).tick ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/store_raft.go:673 kvserver.(*Store).processTick ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:386 kvserver.(*raftSchedulerShard).worker ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:299 kvserver.(*raftScheduler).Start.func2 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 215081427 lock 0xc005eecb08 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 213522676 lock 0xc01966e208 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 210707111 lock 0xc00d5598d0 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? === RUN TestQueryCache/group/exec-and-prepare === PAUSE TestQueryCache/group/exec-and-prepare === CONT TestQueryCache/group/exec-and-prepare === RUN TestQueryCache/group/prepare-hints === PAUSE TestQueryCache/group/prepare-hints === CONT TestQueryCache/group/prepare-hints === RUN TestQueryCache/group/schemachange === PAUSE TestQueryCache/group/schemachange === CONT TestQueryCache/group/schemachange === RUN TestQueryCache/group/simple === PAUSE TestQueryCache/group/simple === CONT TestQueryCache/group/simple === RUN TestQueryCache/group/simple-prepare === PAUSE TestQueryCache/group/simple-prepare === CONT TestQueryCache/group/simple-prepare ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestQueryCache.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
sql: TestQueryCache failed - sql.TestQueryCache [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9357134?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9357134?buildTab=artifacts#/) on release-23.1 @ [7e72aae900c3ff4b44f1643c2d7ba55fbb2cbe23](https://github.com/cockroachdb/cockroach/commits/7e72aae900c3ff4b44f1643c2d7ba55fbb2cbe23): ``` goroutine 204895147 lock 0xc0037d79d0 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:1186 kvserver.(*Replica).tick ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:1185 kvserver.(*Replica).tick ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/store_raft.go:673 kvserver.(*Store).processTick ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:386 kvserver.(*raftSchedulerShard).worker ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:299 kvserver.(*raftScheduler).Start.func2 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 215081427 lock 0xc005eecb08 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 213522676 lock 0xc01966e208 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? goroutine 210707111 lock 0xc00d5598d0 github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:77 kvserver.(*lockedRangefeedStream).Send ??? <<<<< github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_rangefeed.go:76 kvserver.(*lockedRangefeedStream).Send ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:330 rangefeed.(*registration).outputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:351 rangefeed.(*registration).runOutputLoop ??? github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/processor.go:321 rangefeed.(*Processor).run.func1 ??? github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 stop.(*Stopper).RunAsyncTaskEx.func2 ??? === RUN TestQueryCache/group/exec-and-prepare === PAUSE TestQueryCache/group/exec-and-prepare === CONT TestQueryCache/group/exec-and-prepare === RUN TestQueryCache/group/prepare-hints === PAUSE TestQueryCache/group/prepare-hints === CONT TestQueryCache/group/prepare-hints === RUN TestQueryCache/group/schemachange === PAUSE TestQueryCache/group/schemachange === CONT TestQueryCache/group/schemachange === RUN TestQueryCache/group/simple === PAUSE TestQueryCache/group/simple === CONT TestQueryCache/group/simple === RUN TestQueryCache/group/simple-prepare === PAUSE TestQueryCache/group/simple-prepare === CONT TestQueryCache/group/simple-prepare ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestQueryCache.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
sql testquerycache failed sql testquerycache with on release goroutine lock github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kvserver replica tick github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kvserver replica tick github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver store raft go kvserver store processtick github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver scheduler go kvserver raftschedulershard worker github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver scheduler go kvserver raftscheduler start github com cockroachdb cockroach pkg util stop stopper go stop stopper runasynctaskex goroutine lock github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration outputloop github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration runoutputloop github com cockroachdb cockroach pkg kv kvserver rangefeed processor go rangefeed processor run github com cockroachdb cockroach pkg util stop stopper go stop stopper runasynctaskex goroutine lock github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration outputloop github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration runoutputloop github com cockroachdb cockroach pkg kv kvserver rangefeed processor go rangefeed processor run github com cockroachdb cockroach pkg util stop stopper go stop stopper runasynctaskex goroutine lock github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica rangefeed go kvserver lockedrangefeedstream send github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration outputloop github com cockroachdb cockroach pkg kv kvserver rangefeed registry go rangefeed registration runoutputloop github com cockroachdb cockroach pkg kv kvserver rangefeed processor go rangefeed processor run github com cockroachdb cockroach pkg util stop stopper go stop stopper runasynctaskex run testquerycache group exec and prepare pause testquerycache group exec and prepare cont testquerycache group exec and prepare run testquerycache group prepare hints pause testquerycache group prepare hints cont testquerycache group prepare hints run testquerycache group schemachange pause testquerycache group schemachange cont testquerycache group schemachange run testquerycache group simple pause testquerycache group simple cont testquerycache group simple run testquerycache group simple prepare pause testquerycache group simple prepare cont testquerycache group simple prepare parameters tags bazel gss deadlock help see also cc cockroachdb sql queries
0
274,293
29,992,824,381
IssuesEvent
2023-06-26 01:06:42
panasalap/linux-4.19.72_1
https://api.github.com/repos/panasalap/linux-4.19.72_1
opened
CVE-2023-35788 (High) detected in linux-yoctov5.4.51
Mend: dependency security vulnerability
## CVE-2023-35788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/cls_flower.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/cls_flower.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in fl_set_geneve_opt in net/sched/cls_flower.c in the Linux kernel before 6.3.7. It allows an out-of-bounds write in the flower classifier code via TCA_FLOWER_KEY_ENC_OPTS_GENEVE packets. This may result in denial of service or privilege escalation. <p>Publish Date: 2023-06-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-35788>CVE-2023-35788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-35788">https://www.cve.org/CVERecord?id=CVE-2023-35788</a></p> <p>Release Date: 2023-06-16</p> <p>Fix Resolution: v4.19.285,v5.4.246,v5.10.183,v5.15.116,v6.1.33</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-35788 (High) detected in linux-yoctov5.4.51 - ## CVE-2023-35788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/cls_flower.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sched/cls_flower.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in fl_set_geneve_opt in net/sched/cls_flower.c in the Linux kernel before 6.3.7. It allows an out-of-bounds write in the flower classifier code via TCA_FLOWER_KEY_ENC_OPTS_GENEVE packets. This may result in denial of service or privilege escalation. <p>Publish Date: 2023-06-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-35788>CVE-2023-35788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-35788">https://www.cve.org/CVERecord?id=CVE-2023-35788</a></p> <p>Release Date: 2023-06-16</p> <p>Fix Resolution: v4.19.285,v5.4.246,v5.10.183,v5.15.116,v6.1.33</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in base branch master vulnerable source files net sched cls flower c net sched cls flower c vulnerability details an issue was discovered in fl set geneve opt in net sched cls flower c in the linux kernel before it allows an out of bounds write in the flower classifier code via tca flower key enc opts geneve packets this may result in denial of service or privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
9,048
12,130,108,048
IssuesEvent
2020-04-23 00:30:41
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
remove gcp-devrel-py-tools from appengine/standard/blobstore/gcs/requirements-test.txt
priority: p2 remove-gcp-devrel-py-tools type: process
remove gcp-devrel-py-tools from appengine/standard/blobstore/gcs/requirements-test.txt
1.0
remove gcp-devrel-py-tools from appengine/standard/blobstore/gcs/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/blobstore/gcs/requirements-test.txt
process
remove gcp devrel py tools from appengine standard blobstore gcs requirements test txt remove gcp devrel py tools from appengine standard blobstore gcs requirements test txt
1
151,455
13,402,291,899
IssuesEvent
2020-09-03 18:43:20
dogsheep/dogsheep-beta
https://api.github.com/repos/dogsheep/dogsheep-beta
reopened
Mechanism for defining custom display of results
documentation enhancement
Part of #3 - in particular I want to make sure my photos are displayed with a thumbnail.
1.0
Mechanism for defining custom display of results - Part of #3 - in particular I want to make sure my photos are displayed with a thumbnail.
non_process
mechanism for defining custom display of results part of in particular i want to make sure my photos are displayed with a thumbnail
0
590,377
17,777,061,915
IssuesEvent
2021-08-30 20:42:19
SkriptLang/Skript
https://api.github.com/repos/SkriptLang/Skript
closed
If the hex color tag is set to a variable, it does not work.
bug priority: low completed
### Skript/Server Version [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript [Skript] Server Version: git-Paper-127 (MC: 1.17.1) [Skript] Skript Version: 2.6-beta2-nightly-5a6633f [Skript] Installed Skript Addons: ### Bug Description If the hex color tag is set to a variable, it does not work. It worked fine in the previous version. and It doesn't work in latest version. The latest nightly version doesn't work either. so report it ### Expected Behavior As well as variables, using the 'colored' expression does not work properly. ### Steps to Reproduce ``` command /chat.color: trigger: send "<##00ff00>█" send colored "<##00ff00>█" set {_value} to "<##00ff00>█" send {_value} send colored {_value} send uncolored {_value} ``` ### Errors or Screenshots ![2021-08-14_15 55 11](https://user-images.githubusercontent.com/59043053/129438293-779862eb-62d7-4189-9990-5c755e74d97a.png) ### Other _No response_ ### Agreement - [X] I have read the guidelines above and confirm I am following them with this report.
1.0
If the hex color tag is set to a variable, it does not work. - ### Skript/Server Version [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript [Skript] Server Version: git-Paper-127 (MC: 1.17.1) [Skript] Skript Version: 2.6-beta2-nightly-5a6633f [Skript] Installed Skript Addons: ### Bug Description If the hex color tag is set to a variable, it does not work. It worked fine in the previous version. and It doesn't work in latest version. The latest nightly version doesn't work either. so report it ### Expected Behavior As well as variables, using the 'colored' expression does not work properly. ### Steps to Reproduce ``` command /chat.color: trigger: send "<##00ff00>█" send colored "<##00ff00>█" set {_value} to "<##00ff00>█" send {_value} send colored {_value} send uncolored {_value} ``` ### Errors or Screenshots ![2021-08-14_15 55 11](https://user-images.githubusercontent.com/59043053/129438293-779862eb-62d7-4189-9990-5c755e74d97a.png) ### Other _No response_ ### Agreement - [X] I have read the guidelines above and confirm I am following them with this report.
non_process
if the hex color tag is set to a variable it does not work skript server version skript s aliases can be found here skript s documentation can be found here server version git paper mc skript version nightly installed skript addons bug description if the hex color tag is set to a variable it does not work it worked fine in the previous version and it doesn t work in latest version the latest nightly version doesn t work either so report it expected behavior as well as variables using the colored expression does not work properly steps to reproduce command chat color trigger send █ send colored █ set value to █ send value send colored value send uncolored value errors or screenshots other no response agreement i have read the guidelines above and confirm i am following them with this report
0
14,896
18,291,214,867
IssuesEvent
2021-10-05 15:26:05
kcp-dev/kcp
https://api.github.com/repos/kcp-dev/kcp
opened
Create a first sketch of APIs / CRDs for resources like `Workspace`, `WorkspaceShard`, `APIBindings`, etc ...
in-process
Create a first sketch of APIs / CRDs for resources like `Workspace`, `WorkspaceShard`, `APIBindings`, etc ...
1.0
Create a first sketch of APIs / CRDs for resources like `Workspace`, `WorkspaceShard`, `APIBindings`, etc ... - Create a first sketch of APIs / CRDs for resources like `Workspace`, `WorkspaceShard`, `APIBindings`, etc ...
process
create a first sketch of apis crds for resources like workspace workspaceshard apibindings etc create a first sketch of apis crds for resources like workspace workspaceshard apibindings etc
1
12,320
14,879,445,838
IssuesEvent
2021-01-20 07:42:02
lutraconsulting/qgis-crayfish-plugin
https://api.github.com/repos/lutraconsulting/qgis-crayfish-plugin
closed
Create contour from color ramp
enhancement processing
This feature was available in crayfish 2.x: the color ramp from the active dataset can be used for the contouring option.
1.0
Create contour from color ramp - This feature was available in crayfish 2.x: the color ramp from the active dataset can be used for the contouring option.
process
create contour from color ramp this feature was available in crayfish x the color ramp from the active dataset can be used for the contouring option
1
7,788
10,928,178,315
IssuesEvent
2019-11-22 18:25:43
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
PubSub: Add more system tests covering various RBAC-related scenarios
api: pubsub testing type: process
A [regression](https://github.com/googleapis/google-cloud-python/issues/9339) introduced not so long ago could have been prevented if we had more RBAC-related system tests. A system test covering the fix was added in #9507. Since the PubSub backend defines several different [user roles](https://cloud.google.com/pubsub/docs/access-control#roles), system tests must be added to cover at least most common use case scenarios that can be affected by the user role used in the session.
1.0
PubSub: Add more system tests covering various RBAC-related scenarios - A [regression](https://github.com/googleapis/google-cloud-python/issues/9339) introduced not so long ago could have been prevented if we had more RBAC-related system tests. A system test covering the fix was added in #9507. Since the PubSub backend defines several different [user roles](https://cloud.google.com/pubsub/docs/access-control#roles), system tests must be added to cover at least most common use case scenarios that can be affected by the user role used in the session.
process
pubsub add more system tests covering various rbac related scenarios a introduced not so long ago could have been prevented if we had more rbac related system tests a system test covering the fix was added in since the pubsub backend defines several different system tests must be added to cover at least most common use case scenarios that can be affected by the user role used in the session
1
268,807
20,361,840,345
IssuesEvent
2022-02-20 19:56:46
MiguelDuba/git_flow_practice
https://api.github.com/repos/MiguelDuba/git_flow_practice
closed
Un commit que no sigue la convención de código o arreglo a realizar
documentation
La convención del mensaje del último commit no es la esperada: `FIX1: Arreglo pagina arroz con coco` Recuerde que debe tener el siguiente formato: `<Identificador de la corrección>: <Comentario>` Para realizar la corrección del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f` Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
1.0
Un commit que no sigue la convención de código o arreglo a realizar - La convención del mensaje del último commit no es la esperada: `FIX1: Arreglo pagina arroz con coco` Recuerde que debe tener el siguiente formato: `<Identificador de la corrección>: <Comentario>` Para realizar la corrección del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f` Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
non_process
un commit que no sigue la convención de código o arreglo a realizar la convención del mensaje del último commit no es la esperada arreglo pagina arroz con coco recuerde que debe tener el siguiente formato para realizar la corrección del mensaje de commit ejecute los comandos git commit amend y git push f este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado
0
12,595
14,992,981,034
IssuesEvent
2021-01-29 10:38:33
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Default study thumbnail image is not displayed in sites and studies tab
Bug P2 Process: Fixed Process: Tested dev Unknown backend
Steps: 1. Login to SB 2. Don't upload any thumbnail image i.e let be alternate image 3. Launch the study 4. Observe the study in PM Actual: Default study thumbnail image is not displayed in sites and studies tab Expected: Default study thumbnail image should be displayed in sites and studies tab ![Screenshot_3](https://user-images.githubusercontent.com/60386291/104925016-415ce080-59c4-11eb-9b07-4eee4420be45.png)
2.0
[PM] Default study thumbnail image is not displayed in sites and studies tab - Steps: 1. Login to SB 2. Don't upload any thumbnail image i.e let be alternate image 3. Launch the study 4. Observe the study in PM Actual: Default study thumbnail image is not displayed in sites and studies tab Expected: Default study thumbnail image should be displayed in sites and studies tab ![Screenshot_3](https://user-images.githubusercontent.com/60386291/104925016-415ce080-59c4-11eb-9b07-4eee4420be45.png)
process
default study thumbnail image is not displayed in sites and studies tab steps login to sb don t upload any thumbnail image i e let be alternate image launch the study observe the study in pm actual default study thumbnail image is not displayed in sites and studies tab expected default study thumbnail image should be displayed in sites and studies tab
1
7,556
18,241,771,012
IssuesEvent
2021-10-01 13:43:36
dfds/backstage
https://api.github.com/repos/dfds/backstage
closed
Distributed Systems Design - Workshop
Architecture
- [x] Meet with LearnSome - [x] Draft ppt - [x] Draft learnsome profile - [x] Get feedback on LearnSome profile + PPT from Jakob F, Jan W & Martin O - [x] Demo
1.0
Distributed Systems Design - Workshop - - [x] Meet with LearnSome - [x] Draft ppt - [x] Draft learnsome profile - [x] Get feedback on LearnSome profile + PPT from Jakob F, Jan W & Martin O - [x] Demo
non_process
distributed systems design workshop meet with learnsome draft ppt draft learnsome profile get feedback on learnsome profile ppt from jakob f jan w martin o demo
0
3,217
6,277,231,688
IssuesEvent
2017-07-18 11:40:24
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Test: System.Diagnostics.Tests.ProcessTests/TestExitTime failed with "TestExitTime is incorrect"
area-System.Diagnostics.Process os-linux test-run-core
Opened on behalf of @Jiayili1 The test `System.Diagnostics.Tests.ProcessTests/TestExitTime` has failed. TestExitTime is incorrect. TimeBeforeStart 7/13/17 12:51:12 AM Ticks=636355038728165136, ExitTime=7/13/17 12:51:12 AM, Ticks=636355038727555168, ExitTimeUniversal 7/13/17 12:51:12 AM Ticks=636355038727555168, NowUniversal 7/13/17 12:51:12 AM Ticks=636355038727559263 Expected: True Actual: False Stack Trace: at System.Diagnostics.Tests.ProcessTests.TestExitTime() in /root/corefx/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 149 Build : 2.0.0 - 20170713.01 (Core Tests) Failing configurations: - Ubuntu.1610.Amd64-x64 - Debug Detail: https://mc.dot.net/#/product/netcore/200/source/official~2Fcorefx~2Frelease~2F2.0.0~2F/type/test~2Ffunctional~2Fcli~2F/build/20170713.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestExitTime
1.0
Test: System.Diagnostics.Tests.ProcessTests/TestExitTime failed with "TestExitTime is incorrect" - Opened on behalf of @Jiayili1 The test `System.Diagnostics.Tests.ProcessTests/TestExitTime` has failed. TestExitTime is incorrect. TimeBeforeStart 7/13/17 12:51:12 AM Ticks=636355038728165136, ExitTime=7/13/17 12:51:12 AM, Ticks=636355038727555168, ExitTimeUniversal 7/13/17 12:51:12 AM Ticks=636355038727555168, NowUniversal 7/13/17 12:51:12 AM Ticks=636355038727559263 Expected: True Actual: False Stack Trace: at System.Diagnostics.Tests.ProcessTests.TestExitTime() in /root/corefx/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 149 Build : 2.0.0 - 20170713.01 (Core Tests) Failing configurations: - Ubuntu.1610.Amd64-x64 - Debug Detail: https://mc.dot.net/#/product/netcore/200/source/official~2Fcorefx~2Frelease~2F2.0.0~2F/type/test~2Ffunctional~2Fcli~2F/build/20170713.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestExitTime
process
test system diagnostics tests processtests testexittime failed with testexittime is incorrect opened on behalf of the test system diagnostics tests processtests testexittime has failed testexittime is incorrect timebeforestart am ticks exittime am ticks exittimeuniversal am ticks nowuniversal am ticks expected true actual false stack trace at system diagnostics tests processtests testexittime in root corefx src system diagnostics process tests processtests cs line build core tests failing configurations ubuntu debug detail
1
5,000
7,834,687,700
IssuesEvent
2018-06-16 17:11:49
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Add useful plots and get rid of useless plots on plotgen (Trac #824)
Migrated from Trac bmg2@uwm.edu post_processing task
Plotgen could be improved by adding more useful panels to it. However, that also increases the amount of space that each plotgen ```.maff``` files takes up, and we don't want to go over the attachment limit for Trac. In order to make room for new plots, useless panels should be taken off plotgen. An example of a useless panel would be graupel mixing ratio for the FIRE stratocumulus case. Attachments: [plotgen_test_r8643.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/clubb/plotgen_test_r8643.maff) [plotgen_test_r8655.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/clubb/plotgen_test_r8655.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/824 ```json { "status": "closed", "changetime": "2018-06-05T18:22:00", "description": "Plotgen could be improved by adding more useful panels to it. However, that also increases the amount of space that each plotgen {{{.maff}}} files takes up, and we don't want to go over the attachment limit for Trac. In order to make room for new plots, useless panels should be taken off plotgen. An example of a useless panel would be graupel mixing ratio for the FIRE stratocumulus case.", "reporter": "bmg2@uwm.edu", "cc": "vlarson@uwm.edu", "resolution": "fixed", "_ts": "1528222920626922", "component": "post_processing", "summary": "Add useful plots and get rid of useless plots on plotgen", "priority": "minor", "keywords": "new plots", "time": "2018-04-19T21:46:14", "milestone": "Improve Plotgen", "owner": "bmg2@uwm.edu", "type": "task" } ```
1.0
Add useful plots and get rid of useless plots on plotgen (Trac #824) - Plotgen could be improved by adding more useful panels to it. However, that also increases the amount of space that each plotgen ```.maff``` files takes up, and we don't want to go over the attachment limit for Trac. In order to make room for new plots, useless panels should be taken off plotgen. An example of a useless panel would be graupel mixing ratio for the FIRE stratocumulus case. Attachments: [plotgen_test_r8643.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/clubb/plotgen_test_r8643.maff) [plotgen_test_r8655.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/clubb/plotgen_test_r8655.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/824 ```json { "status": "closed", "changetime": "2018-06-05T18:22:00", "description": "Plotgen could be improved by adding more useful panels to it. However, that also increases the amount of space that each plotgen {{{.maff}}} files takes up, and we don't want to go over the attachment limit for Trac. In order to make room for new plots, useless panels should be taken off plotgen. An example of a useless panel would be graupel mixing ratio for the FIRE stratocumulus case.", "reporter": "bmg2@uwm.edu", "cc": "vlarson@uwm.edu", "resolution": "fixed", "_ts": "1528222920626922", "component": "post_processing", "summary": "Add useful plots and get rid of useless plots on plotgen", "priority": "minor", "keywords": "new plots", "time": "2018-04-19T21:46:14", "milestone": "Improve Plotgen", "owner": "bmg2@uwm.edu", "type": "task" } ```
process
add useful plots and get rid of useless plots on plotgen trac plotgen could be improved by adding more useful panels to it however that also increases the amount of space that each plotgen maff files takes up and we don t want to go over the attachment limit for trac in order to make room for new plots useless panels should be taken off plotgen an example of a useless panel would be graupel mixing ratio for the fire stratocumulus case attachments migrated from json status closed changetime description plotgen could be improved by adding more useful panels to it however that also increases the amount of space that each plotgen maff files takes up and we don t want to go over the attachment limit for trac in order to make room for new plots useless panels should be taken off plotgen an example of a useless panel would be graupel mixing ratio for the fire stratocumulus case reporter uwm edu cc vlarson uwm edu resolution fixed ts component post processing summary add useful plots and get rid of useless plots on plotgen priority minor keywords new plots time milestone improve plotgen owner uwm edu type task
1
591,010
17,792,961,447
IssuesEvent
2021-08-31 18:26:05
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[rv_dm] Current life cycle gating incomplete
Priority:P0 Type:Bug
The current rv_dm life cycle gating does not gate on the input of the JTAG->DMI interface. This means certain jtag functionality, such as ID, bypass etc are still available even though the DMI interface is no longer useable. This makes it a bit more difficult for DV to verify, so we should probably consider updating this.
1.0
[rv_dm] Current life cycle gating incomplete - The current rv_dm life cycle gating does not gate on the input of the JTAG->DMI interface. This means certain jtag functionality, such as ID, bypass etc are still available even though the DMI interface is no longer useable. This makes it a bit more difficult for DV to verify, so we should probably consider updating this.
non_process
current life cycle gating incomplete the current rv dm life cycle gating does not gate on the input of the jtag dmi interface this means certain jtag functionality such as id bypass etc are still available even though the dmi interface is no longer useable this makes it a bit more difficult for dv to verify so we should probably consider updating this
0
8,801
11,908,262,183
IssuesEvent
2020-03-31 00:26:32
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing algorithms are not able to add the output layer to an exiting GPKG or SQLite container without completely overwrite it
Bug Processing
Author Name: **Andrea Giudiceandrea** (@agiudiceandrea) Original Redmine Issue: [20026](https://issues.qgis.org/issues/20026) Affected QGIS version: 3.4.0 Redmine category:processing/core --- Core processing algorithms are not able to add the output layer to an exiting SQLite (maybe others?) container file without completely overwrite it. GDAL/OGR processing algorithms are not able to do it even with GPKG geopackage file.
1.0
Processing algorithms are not able to add the output layer to an exiting GPKG or SQLite container without completely overwrite it - Author Name: **Andrea Giudiceandrea** (@agiudiceandrea) Original Redmine Issue: [20026](https://issues.qgis.org/issues/20026) Affected QGIS version: 3.4.0 Redmine category:processing/core --- Core processing algorithms are not able to add the output layer to an exiting SQLite (maybe others?) container file without completely overwrite it. GDAL/OGR processing algorithms are not able to do it even with GPKG geopackage file.
process
processing algorithms are not able to add the output layer to an exiting gpkg or sqlite container without completely overwrite it author name andrea giudiceandrea agiudiceandrea original redmine issue affected qgis version redmine category processing core core processing algorithms are not able to add the output layer to an exiting sqlite maybe others container file without completely overwrite it gdal ogr processing algorithms are not able to do it even with gpkg geopackage file
1
40,271
2,868,312,639
IssuesEvent
2015-06-05 18:05:47
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Console: refresh is sometimes needed after starting build
component/web kind/bug priority/P2
Sometimes when you trigger a build from the console, the UI doesn't show the build and you have to refresh in order to see it.
1.0
Console: refresh is sometimes needed after starting build - Sometimes when you trigger a build from the console, the UI doesn't show the build and you have to refresh in order to see it.
non_process
console refresh is sometimes needed after starting build sometimes when you trigger a build from the console the ui doesn t show the build and you have to refresh in order to see it
0
296,420
9,115,465,861
IssuesEvent
2019-02-22 05:07:54
WeAreDevs/material-clicker
https://api.github.com/repos/WeAreDevs/material-clicker
opened
Achievements
enhancement priority: low
When doing certain things, you can trigger an achievement. I propose an API Something similar to promises. ```js registerAchievement({ id: 'win-universe', name: 'Win the Universe' description: 'Example to show how cool the world is.' }, (award) => { // add event handles and stuff to detect if they get the achievement award(); // call award to give the achievement. }); ```
1.0
Achievements - When doing certain things, you can trigger an achievement. I propose an API Something similar to promises. ```js registerAchievement({ id: 'win-universe', name: 'Win the Universe' description: 'Example to show how cool the world is.' }, (award) => { // add event handles and stuff to detect if they get the achievement award(); // call award to give the achievement. }); ```
non_process
achievements when doing certain things you can trigger an achievement i propose an api something similar to promises js registerachievement id win universe name win the universe description example to show how cool the world is award add event handles and stuff to detect if they get the achievement award call award to give the achievement
0
77,176
9,982,611,364
IssuesEvent
2019-07-10 10:15:19
tomusdrw/rust-web3
https://api.github.com/repos/tomusdrw/rust-web3
closed
[Question] Regarding Example
documentation
Hello, I am wondering if perhaps the readme for example is missing something? I have cloned the project and followed the guide from `example/readme.md`. That is ``` ganache-cli -m "hamster coin cup brief quote trick stove draft hobby strong caught unable" cargo run --example contract ``` Everything appears to compile fine, but the only thing appearing in ganache is a revert statement: ``` eth_sendTransaction Transaction: 0x21860b87e3f3085dd78c659042af9d214f3c8ce449a7886815c3b99fc8795390 Contract created: 0x4d470146215d085c75767b717dbb8d3b4468f893 Gas usage: 241727 Block Number: 1 Block Time: Wed Jun 26 2019 17:19:40 GMT+0200 (Central European Summer Time) Runtime Error: revert eth_newBlockFilter eth_getFilterChanges ``` The `cargo run` itself just appears to freeze/hang like so; ``` Finished dev [unoptimized + debuginfo] target(s) in 2m 12s Running `target/debug/examples/contract` ``` I am still kinda new to rust, but have tried this with both rust versions `1.33.0` and `1.34.0`. After including a few print statements, it appears to be a hanging at `Contract::deploy` https://github.com/tomusdrw/rust-web3/blob/21cb38c7203a026f292d65059272b88e2050626e/examples/contract.rs#L18-L33 Not sure if I am doing something wrong or if there might be something out of date here...
1.0
[Question] Regarding Example - Hello, I am wondering if perhaps the readme for example is missing something? I have cloned the project and followed the guide from `example/readme.md`. That is ``` ganache-cli -m "hamster coin cup brief quote trick stove draft hobby strong caught unable" cargo run --example contract ``` Everything appears to compile fine, but the only thing appearing in ganache is a revert statement: ``` eth_sendTransaction Transaction: 0x21860b87e3f3085dd78c659042af9d214f3c8ce449a7886815c3b99fc8795390 Contract created: 0x4d470146215d085c75767b717dbb8d3b4468f893 Gas usage: 241727 Block Number: 1 Block Time: Wed Jun 26 2019 17:19:40 GMT+0200 (Central European Summer Time) Runtime Error: revert eth_newBlockFilter eth_getFilterChanges ``` The `cargo run` itself just appears to freeze/hang like so; ``` Finished dev [unoptimized + debuginfo] target(s) in 2m 12s Running `target/debug/examples/contract` ``` I am still kinda new to rust, but have tried this with both rust versions `1.33.0` and `1.34.0`. After including a few print statements, it appears to be a hanging at `Contract::deploy` https://github.com/tomusdrw/rust-web3/blob/21cb38c7203a026f292d65059272b88e2050626e/examples/contract.rs#L18-L33 Not sure if I am doing something wrong or if there might be something out of date here...
non_process
regarding example hello i am wondering if perhaps the readme for example is missing something i have cloned the project and followed the guide from example readme md that is ganache cli m hamster coin cup brief quote trick stove draft hobby strong caught unable cargo run example contract everything appears to compile fine but the only thing appearing in ganache is a revert statement eth sendtransaction transaction contract created gas usage block number block time wed jun gmt central european summer time runtime error revert eth newblockfilter eth getfilterchanges the cargo run itself just appears to freeze hang like so finished dev target s in running target debug examples contract i am still kinda new to rust but have tried this with both rust versions and after including a few print statements it appears to be a hanging at contract deploy not sure if i am doing something wrong or if there might be something out of date here
0
22,686
31,987,992,167
IssuesEvent
2023-09-21 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 21 Sep 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### GL-Fusion: Global-Local Fusion Network for Multi-view Echocardiogram Video Segmentation - **Authors:** Ziyang Zheng, Jiewen Yang, Xinpeng Ding, Xiaowei Xu, Xiaomeng Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.11144 - **Pdf link:** https://arxiv.org/pdf/2309.11144 - **Abstract** Cardiac structure segmentation from echocardiogram videos plays a crucial role in diagnosing heart disease. The combination of multi-view echocardiogram data is essential to enhance the accuracy and robustness of automated methods. However, due to the visual disparity of the data, deriving cross-view context information remains a challenging task, and unsophisticated fusion strategies can even lower performance. In this study, we propose a novel Gobal-Local fusion (GL-Fusion) network to jointly utilize multi-view information globally and locally that improve the accuracy of echocardiogram analysis. Specifically, a Multi-view Global-based Fusion Module (MGFM) is proposed to extract global context information and to explore the cyclic relationship of different heartbeat cycles in an echocardiogram video. Additionally, a Multi-view Local-based Fusion Module (MLFM) is designed to extract correlations of cardiac structures from different views. Furthermore, we collect a multi-view echocardiogram video dataset (MvEVD) to evaluate our method. Our method achieves an 82.29% average dice score, which demonstrates a 7.83% improvement over the baseline method, and outperforms other existing state-of-the-art methods. To our knowledge, this is the first exploration of a multi-view method for echocardiogram video segmentation. Code available at: https://github.com/xmed-lab/GL-Fusion ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### DreamLLM: Synergistic Multimodal Comprehension and Creation - **Authors:** Runpei Dong, Chunrui Han, Yuang Peng, Zekun Qi, Zheng Ge, Jinrong Yang, Liang Zhao, Jianjian Sun, Hongyu Zhou, Haoran Wei, Xiangwen Kong, Xiangyu Zhang, Kaisheng Ma, Li Yi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.11499 - **Pdf link:** https://arxiv.org/pdf/2309.11499 - **Abstract** This paper presents DreamLLM, a learning framework that first achieves versatile Multimodal Large Language Models (MLLMs) empowered with frequently overlooked synergy between multimodal comprehension and creation. DreamLLM operates on two fundamental principles. The first focuses on the generative modeling of both language and image posteriors by direct sampling in the raw multimodal space. This approach circumvents the limitations and information loss inherent to external feature extractors like CLIP, and a more thorough multimodal understanding is obtained. Second, DreamLLM fosters the generation of raw, interleaved documents, modeling both text and image contents, along with unstructured layouts. This allows DreamLLM to learn all conditional, marginal, and joint multimodal distributions effectively. As a result, DreamLLM is the first MLLM capable of generating free-form interleaved content. Comprehensive experiments highlight DreamLLM's superior performance as a zero-shot multimodal generalist, reaping from the enhanced learning synergy. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 21 Sep 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### GL-Fusion: Global-Local Fusion Network for Multi-view Echocardiogram Video Segmentation - **Authors:** Ziyang Zheng, Jiewen Yang, Xinpeng Ding, Xiaowei Xu, Xiaomeng Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.11144 - **Pdf link:** https://arxiv.org/pdf/2309.11144 - **Abstract** Cardiac structure segmentation from echocardiogram videos plays a crucial role in diagnosing heart disease. The combination of multi-view echocardiogram data is essential to enhance the accuracy and robustness of automated methods. However, due to the visual disparity of the data, deriving cross-view context information remains a challenging task, and unsophisticated fusion strategies can even lower performance. In this study, we propose a novel Gobal-Local fusion (GL-Fusion) network to jointly utilize multi-view information globally and locally that improve the accuracy of echocardiogram analysis. Specifically, a Multi-view Global-based Fusion Module (MGFM) is proposed to extract global context information and to explore the cyclic relationship of different heartbeat cycles in an echocardiogram video. Additionally, a Multi-view Local-based Fusion Module (MLFM) is designed to extract correlations of cardiac structures from different views. Furthermore, we collect a multi-view echocardiogram video dataset (MvEVD) to evaluate our method. Our method achieves an 82.29% average dice score, which demonstrates a 7.83% improvement over the baseline method, and outperforms other existing state-of-the-art methods. To our knowledge, this is the first exploration of a multi-view method for echocardiogram video segmentation. Code available at: https://github.com/xmed-lab/GL-Fusion ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### DreamLLM: Synergistic Multimodal Comprehension and Creation - **Authors:** Runpei Dong, Chunrui Han, Yuang Peng, Zekun Qi, Zheng Ge, Jinrong Yang, Liang Zhao, Jianjian Sun, Hongyu Zhou, Haoran Wei, Xiangwen Kong, Xiangyu Zhang, Kaisheng Ma, Li Yi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.11499 - **Pdf link:** https://arxiv.org/pdf/2309.11499 - **Abstract** This paper presents DreamLLM, a learning framework that first achieves versatile Multimodal Large Language Models (MLLMs) empowered with frequently overlooked synergy between multimodal comprehension and creation. DreamLLM operates on two fundamental principles. The first focuses on the generative modeling of both language and image posteriors by direct sampling in the raw multimodal space. This approach circumvents the limitations and information loss inherent to external feature extractors like CLIP, and a more thorough multimodal understanding is obtained. Second, DreamLLM fosters the generation of raw, interleaved documents, modeling both text and image contents, along with unstructured layouts. This allows DreamLLM to learn all conditional, marginal, and joint multimodal distributions effectively. As a result, DreamLLM is the first MLLM capable of generating free-form interleaved content. Comprehensive experiments highlight DreamLLM's superior performance as a zero-shot multimodal generalist, reaping from the enhanced learning synergy. ## Keyword: raw image There is no result
process
new submissions for thu sep keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp gl fusion global local fusion network for multi view echocardiogram video segmentation authors ziyang zheng jiewen yang xinpeng ding xiaowei xu xiaomeng li subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract cardiac structure segmentation from echocardiogram videos plays a crucial role in diagnosing heart disease the combination of multi view echocardiogram data is essential to enhance the accuracy and robustness of automated methods however due to the visual disparity of the data deriving cross view context information remains a challenging task and unsophisticated fusion strategies can even lower performance in this study we propose a novel gobal local fusion gl fusion network to jointly utilize multi view information globally and locally that improve the accuracy of echocardiogram analysis specifically a multi view global based fusion module mgfm is proposed to extract global context information and to explore the cyclic relationship of different heartbeat cycles in an echocardiogram video additionally a multi view local based fusion module mlfm is designed to extract correlations of cardiac structures from different views furthermore we collect a multi view echocardiogram video dataset mvevd to evaluate our method our method achieves an average dice score which demonstrates a improvement over the baseline method and outperforms other existing state of the art methods to our knowledge this is the first exploration of a multi view method for echocardiogram video segmentation code available at keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw dreamllm synergistic multimodal comprehension and creation authors runpei dong chunrui han yuang peng zekun qi zheng ge jinrong yang liang zhao jianjian sun hongyu zhou haoran wei xiangwen kong xiangyu zhang kaisheng ma li yi subjects computer vision and pattern recognition cs cv computation and language cs cl machine learning cs lg arxiv link pdf link abstract this paper presents dreamllm a learning framework that first achieves versatile multimodal large language models mllms empowered with frequently overlooked synergy between multimodal comprehension and creation dreamllm operates on two fundamental principles the first focuses on the generative modeling of both language and image posteriors by direct sampling in the raw multimodal space this approach circumvents the limitations and information loss inherent to external feature extractors like clip and a more thorough multimodal understanding is obtained second dreamllm fosters the generation of raw interleaved documents modeling both text and image contents along with unstructured layouts this allows dreamllm to learn all conditional marginal and joint multimodal distributions effectively as a result dreamllm is the first mllm capable of generating free form interleaved content comprehensive experiments highlight dreamllm s superior performance as a zero shot multimodal generalist reaping from the enhanced learning synergy keyword raw image there is no result
1
62,773
12,240,559,054
IssuesEvent
2020-05-05 00:44:36
tinymce/tinymce
https://api.github.com/repos/tinymce/tinymce
closed
deleting in lists does not work properally if <br> tag is removed with "code" or "paste" plugins
plugin: code status: verified type: bug
If you type out a list in an editor: ``` 1. Test 2. Test 3. Test 4. Test 5. Test ``` then using the "_code_" plugin to view the source and click ok to apply changes then start pressing the **delete** button (at least 3 times) from the end of "_1. Test_" it will pull the next sibling element up to soon... and if you keep pressing delete it will continue to pull others early as well. --- so far I have found that normally each `<li>` or any tag with text for that matter will have a `<br data-mce-bogus="1">` after the text. In the "_code_" plugin `win.find('#code').value(editor.getContent({source_view: true}));` strips out the `<br>` tags for viewing in the modal and does not put them back when it sets the contents back to the editor. I also notice that if you paste content with a list in from word or some other source it will also not add the `<br>` tags to the markup during the paste processing.... --- So either the deleting needs to not care that the `<br>` tags are no longer there, maybe the bookmarks needs updated? as it seems to be one row ahead? or it needs something to check and add them back in if missing/removed --- As a work around I have been able to manually add the `<br>` tags back in with the following code but is less than ideal.... ``` var missingBrTags = $('#editor iframe').contents().find('.mce-content-body').find('li,p,div,span').not(':has(>br)'); $.each(missingBrTags, function(missingBrTagIndex, missingBrTag) { if ($(missingBrTag).children().length > 0) { // contains more that just text... var firstNestedDomElement = $(missingBrTag).children('ol,ul,p,div,span').first(); if (firstNestedDomElement.length > 0 ) { // contains major dom elements insert before the first one $('<br data-mce-bogus="1">').insertBefore(firstNestedDomElement); } else { // only has minor markup tags like <strong> or <em>... put br after them $(missingBrTag).append('<br data-mce-bogus="1">'); } } else { // contains only text $(missingBrTag).append('<br data-mce-bogus="1">'); } }); ``` using tinymce 4.4.0 & IE 11 or the editor on the main page of https://www.tinymce.com/
1.0
deleting in lists does not work properally if <br> tag is removed with "code" or "paste" plugins - If you type out a list in an editor: ``` 1. Test 2. Test 3. Test 4. Test 5. Test ``` then using the "_code_" plugin to view the source and click ok to apply changes then start pressing the **delete** button (at least 3 times) from the end of "_1. Test_" it will pull the next sibling element up to soon... and if you keep pressing delete it will continue to pull others early as well. --- so far I have found that normally each `<li>` or any tag with text for that matter will have a `<br data-mce-bogus="1">` after the text. In the "_code_" plugin `win.find('#code').value(editor.getContent({source_view: true}));` strips out the `<br>` tags for viewing in the modal and does not put them back when it sets the contents back to the editor. I also notice that if you paste content with a list in from word or some other source it will also not add the `<br>` tags to the markup during the paste processing.... --- So either the deleting needs to not care that the `<br>` tags are no longer there, maybe the bookmarks needs updated? as it seems to be one row ahead? or it needs something to check and add them back in if missing/removed --- As a work around I have been able to manually add the `<br>` tags back in with the following code but is less than ideal.... ``` var missingBrTags = $('#editor iframe').contents().find('.mce-content-body').find('li,p,div,span').not(':has(>br)'); $.each(missingBrTags, function(missingBrTagIndex, missingBrTag) { if ($(missingBrTag).children().length > 0) { // contains more that just text... var firstNestedDomElement = $(missingBrTag).children('ol,ul,p,div,span').first(); if (firstNestedDomElement.length > 0 ) { // contains major dom elements insert before the first one $('<br data-mce-bogus="1">').insertBefore(firstNestedDomElement); } else { // only has minor markup tags like <strong> or <em>... put br after them $(missingBrTag).append('<br data-mce-bogus="1">'); } } else { // contains only text $(missingBrTag).append('<br data-mce-bogus="1">'); } }); ``` using tinymce 4.4.0 & IE 11 or the editor on the main page of https://www.tinymce.com/
non_process
deleting in lists does not work properally if tag is removed with code or paste plugins if you type out a list in an editor test test test test test then using the code plugin to view the source and click ok to apply changes then start pressing the delete button at least times from the end of test it will pull the next sibling element up to soon and if you keep pressing delete it will continue to pull others early as well so far i have found that normally each or any tag with text for that matter will have a after the text in the code plugin win find code value editor getcontent source view true strips out the tags for viewing in the modal and does not put them back when it sets the contents back to the editor i also notice that if you paste content with a list in from word or some other source it will also not add the tags to the markup during the paste processing so either the deleting needs to not care that the tags are no longer there maybe the bookmarks needs updated as it seems to be one row ahead or it needs something to check and add them back in if missing removed as a work around i have been able to manually add the tags back in with the following code but is less than ideal var missingbrtags editor iframe contents find mce content body find li p div span not has br each missingbrtags function missingbrtagindex missingbrtag if missingbrtag children length contains more that just text var firstnesteddomelement missingbrtag children ol ul p div span first if firstnesteddomelement length contains major dom elements insert before the first one insertbefore firstnesteddomelement else only has minor markup tags like or put br after them missingbrtag append else contains only text missingbrtag append using tinymce ie or the editor on the main page of
0
60,102
3,120,764,896
IssuesEvent
2015-09-05 01:40:28
framingeinstein/issues-test
https://api.github.com/repos/framingeinstein/issues-test
closed
SPK-83: Breadcrumbs: Structure is not correct on About Speakman & Support
priority:normal resolution:fixed type:enhancement
Hi Sam, Creating a ticket for this issue we found last week. Please notice that the structure isn't correct for About Speakman & Support. It seems like Blog is a parent. Thanks, Jessica
1.0
SPK-83: Breadcrumbs: Structure is not correct on About Speakman & Support - Hi Sam, Creating a ticket for this issue we found last week. Please notice that the structure isn't correct for About Speakman & Support. It seems like Blog is a parent. Thanks, Jessica
non_process
spk breadcrumbs structure is not correct on about speakman support hi sam creating a ticket for this issue we found last week please notice that the structure isn t correct for about speakman support it seems like blog is a parent thanks jessica
0
9,942
12,975,530,779
IssuesEvent
2020-07-21 17:09:23
tc39/proposal-promise-any
https://api.github.com/repos/tc39/proposal-promise-any
closed
Advance to stage 4
process
Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] [Test262](https://github.com/tc39/test262) acceptance tests have been written for mainline usage scenarios, and merged - [x] `Promise.any`: https://github.com/tc39/test262/tree/master/test/built-ins/Promise/any - [x] `AggregateError`: https://github.com/tc39/test262/tree/master/test/built-ins/NativeErrors/AggregateError > - [x] Two compatible implementations which pass the acceptance tests https://github.com/tc39/proposal-promise-any#implementations Bug tickets to track: - [x] Chrome/V8: https://bugs.chromium.org/p/v8/issues/detail?id=9808 - [ ] ~~Edge/Chakra: https://github.com/microsoft/ChakraCore/pull/6301~~ - [x] Firefox/SpiderMonkey: https://bugzilla.mozilla.org/show_bug.cgi?id=1568903 - [x] Safari/JavaScriptCore: https://bugs.webkit.org/show_bug.cgi?id=202566 - [x] engine262: https://github.com/engine262/engine262/commit/c68877ef1c4633daac8b58b5ce1876f709c1cc16 - [x] XS: https://blog.moddable.com/blog/xs10/ > - [x] Significant in-the-field experience with shipping implementations, such as that provided by two independent VMs https://github.com/tc39/proposal-promise-any#implementations > - [x] A pull request has been sent to https://github.com/tc39/ecma262 with the integrated spec text https://github.com/tc39/ecma262/pull/2040 > - [ ] The ECMAScript editor has signed off on the pull request TODO
1.0
Advance to stage 4 - Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] [Test262](https://github.com/tc39/test262) acceptance tests have been written for mainline usage scenarios, and merged - [x] `Promise.any`: https://github.com/tc39/test262/tree/master/test/built-ins/Promise/any - [x] `AggregateError`: https://github.com/tc39/test262/tree/master/test/built-ins/NativeErrors/AggregateError > - [x] Two compatible implementations which pass the acceptance tests https://github.com/tc39/proposal-promise-any#implementations Bug tickets to track: - [x] Chrome/V8: https://bugs.chromium.org/p/v8/issues/detail?id=9808 - [ ] ~~Edge/Chakra: https://github.com/microsoft/ChakraCore/pull/6301~~ - [x] Firefox/SpiderMonkey: https://bugzilla.mozilla.org/show_bug.cgi?id=1568903 - [x] Safari/JavaScriptCore: https://bugs.webkit.org/show_bug.cgi?id=202566 - [x] engine262: https://github.com/engine262/engine262/commit/c68877ef1c4633daac8b58b5ce1876f709c1cc16 - [x] XS: https://blog.moddable.com/blog/xs10/ > - [x] Significant in-the-field experience with shipping implementations, such as that provided by two independent VMs https://github.com/tc39/proposal-promise-any#implementations > - [x] A pull request has been sent to https://github.com/tc39/ecma262 with the integrated spec text https://github.com/tc39/ecma262/pull/2040 > - [ ] The ECMAScript editor has signed off on the pull request TODO
process
advance to stage criteria taken from minus those from previous stages acceptance tests have been written for mainline usage scenarios and merged promise any aggregateerror two compatible implementations which pass the acceptance tests bug tickets to track chrome edge chakra firefox spidermonkey safari javascriptcore xs significant in the field experience with shipping implementations such as that provided by two independent vms a pull request has been sent to with the integrated spec text the ecmascript editor has signed off on the pull request todo
1
9,645
12,605,062,111
IssuesEvent
2020-06-11 15:53:41
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Cooling Tower Form Input Ordering
Process Cooling
Feedback from ORNL testing: **Order of Input fields** 1. Group 1: can we do this order: Water Flow Rate, Cooling Load, then Annual Operating hours? Typically, water flow rate is the easiest to get and operators and cooling load and operating hours are a little harder. 2. Group 2: can we have Cycles of Concentration and Drift Eliminator. 3. Group 3: can we have Evaporation Loss Correction Factor and Drift Loss Factor. Users do not have to change the values in this group. Maybe we can give it a different color? {Kristina note: I personally don't think Drift Loss Factor should be that far away from Drift Eliminator. If you think you agree, maybe don't have 3 groups, and have it be "Cycles", "Drift Y/N", "Drift Factor", "Evap Factor". I don't like the idea of "different color" either. }
1.0
Cooling Tower Form Input Ordering - Feedback from ORNL testing: **Order of Input fields** 1. Group 1: can we do this order: Water Flow Rate, Cooling Load, then Annual Operating hours? Typically, water flow rate is the easiest to get and operators and cooling load and operating hours are a little harder. 2. Group 2: can we have Cycles of Concentration and Drift Eliminator. 3. Group 3: can we have Evaporation Loss Correction Factor and Drift Loss Factor. Users do not have to change the values in this group. Maybe we can give it a different color? {Kristina note: I personally don't think Drift Loss Factor should be that far away from Drift Eliminator. If you think you agree, maybe don't have 3 groups, and have it be "Cycles", "Drift Y/N", "Drift Factor", "Evap Factor". I don't like the idea of "different color" either. }
process
cooling tower form input ordering feedback from ornl testing order of input fields group can we do this order water flow rate cooling load then annual operating hours typically water flow rate is the easiest to get and operators and cooling load and operating hours are a little harder group can we have cycles of concentration and drift eliminator group can we have evaporation loss correction factor and drift loss factor users do not have to change the values in this group maybe we can give it a different color kristina note i personally don t think drift loss factor should be that far away from drift eliminator if you think you agree maybe don t have groups and have it be cycles drift y n drift factor evap factor i don t like the idea of different color either
1
13,994
16,766,075,086
IssuesEvent
2021-06-14 08:58:05
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
opened
Masked values when regridding datasets with coord `circular=False`
cmor preprocessor
**Describe the bug** At the BSC, it was reported that when regridding dataset MCM-UA-1-0, a white line with masked values appeared for lon > 355. After some debugging it turns out that when the longitudes are not circular and do not span fully from [0,360], those masked values can appear after regridding. This dataset got fixed in #1098, because (among other issues) the last bound in atmospheric coordinates was not set properly. That issue makes iris set the `circular` attribute to `False`, and makes the dataset come out of the regridding with masked values. Once the bounds are fixed, a quick solution would be to set the circular attribute to `True` in the same fix, since the condition that iris imposes to set this attribute checks for the fixed bound. But I was wondering if it would be ok to check for this attribute, either during the cmor checks or during the regridding, to make a more general fix.
1.0
Masked values when regridding datasets with coord `circular=False` - **Describe the bug** At the BSC, it was reported that when regridding dataset MCM-UA-1-0, a white line with masked values appeared for lon > 355. After some debugging it turns out that when the longitudes are not circular and do not span fully from [0,360], those masked values can appear after regridding. This dataset got fixed in #1098, because (among other issues) the last bound in atmospheric coordinates was not set properly. That issue makes iris set the `circular` attribute to `False`, and makes the dataset come out of the regridding with masked values. Once the bounds are fixed, a quick solution would be to set the circular attribute to `True` in the same fix, since the condition that iris imposes to set this attribute checks for the fixed bound. But I was wondering if it would be ok to check for this attribute, either during the cmor checks or during the regridding, to make a more general fix.
process
masked values when regridding datasets with coord circular false describe the bug at the bsc it was reported that when regridding dataset mcm ua a white line with masked values appeared for lon after some debugging it turns out that when the longitudes are not circular and do not span fully from those masked values can appear after regridding this dataset got fixed in because among other issues the last bound in atmospheric coordinates was not set properly that issue makes iris set the circular attribute to false and makes the dataset come out of the regridding with masked values once the bounds are fixed a quick solution would be to set the circular attribute to true in the same fix since the condition that iris imposes to set this attribute checks for the fixed bound but i was wondering if it would be ok to check for this attribute either during the cmor checks or during the regridding to make a more general fix
1
83,051
3,621,349,284
IssuesEvent
2016-02-08 23:41:08
isolver/OpenHandWrite
https://api.github.com/repos/isolver/OpenHandWrite
closed
Loading source data file fails if .mwp file with same name exists in same folder.
bug MarkWrite Priority 1
If both a source file, like input.txyp, and a markwrite project file, like input.mwp, are in the same folder, trying to load input.txyp fails because _detectAssociatedSegmentTagsFile() function is reading 'input.mwp' as a possible segment tags file and crashing because it is a binary file.
1.0
Loading source data file fails if .mwp file with same name exists in same folder. - If both a source file, like input.txyp, and a markwrite project file, like input.mwp, are in the same folder, trying to load input.txyp fails because _detectAssociatedSegmentTagsFile() function is reading 'input.mwp' as a possible segment tags file and crashing because it is a binary file.
non_process
loading source data file fails if mwp file with same name exists in same folder if both a source file like input txyp and a markwrite project file like input mwp are in the same folder trying to load input txyp fails because detectassociatedsegmenttagsfile function is reading input mwp as a possible segment tags file and crashing because it is a binary file
0
426,688
29,579,935,762
IssuesEvent
2023-06-07 04:27:29
jupyterlab/jupyterlab-desktop
https://api.github.com/repos/jupyterlab/jupyterlab-desktop
closed
venv selection fails
bug documentation
## Description Upon selecting an environment and choosing a venv python one. I am getting this error. ![image](https://user-images.githubusercontent.com/57333254/222949183-78804067-5f54-4f99-9dec-f237691fb90c.png) ## Reproduce Create a virtual environment with venv and then select it.
1.0
venv selection fails - ## Description Upon selecting an environment and choosing a venv python one. I am getting this error. ![image](https://user-images.githubusercontent.com/57333254/222949183-78804067-5f54-4f99-9dec-f237691fb90c.png) ## Reproduce Create a virtual environment with venv and then select it.
non_process
venv selection fails description upon selecting an environment and choosing a venv python one i am getting this error reproduce create a virtual environment with venv and then select it
0
2,577
5,332,789,030
IssuesEvent
2017-02-15 23:03:23
csstree/stylelint-validator
https://api.github.com/repos/csstree/stylelint-validator
closed
sass interpolation
enhancement preprocessors
Slightly related to #4, I've found our codese problems coming from [sass interpolation]( http://sass-lang.com/documentation/file.SASS_REFERENCE.html#interpolation_) ``` .navbar-static-top .dropdown .dropdown-menu { max-height: calc(100vh - #{$navbar-height}); overflow-y: auto; } ``` ``` >> theme/boost/scss/moodle/undo.scss >> 138:5 ✖ Can't parse value "calc(100vh - #{$navbar-height})" csstree/validator ``` Thanks for this project!
1.0
sass interpolation - Slightly related to #4, I've found our codese problems coming from [sass interpolation]( http://sass-lang.com/documentation/file.SASS_REFERENCE.html#interpolation_) ``` .navbar-static-top .dropdown .dropdown-menu { max-height: calc(100vh - #{$navbar-height}); overflow-y: auto; } ``` ``` >> theme/boost/scss/moodle/undo.scss >> 138:5 ✖ Can't parse value "calc(100vh - #{$navbar-height})" csstree/validator ``` Thanks for this project!
process
sass interpolation slightly related to i ve found our codese problems coming from navbar static top dropdown dropdown menu max height calc navbar height overflow y auto theme boost scss moodle undo scss ✖ can t parse value calc navbar height csstree validator thanks for this project
1
141,889
12,991,206,642
IssuesEvent
2020-07-23 02:47:06
KuChainNetwork/Project-Decalogue
https://api.github.com/repos/KuChainNetwork/Project-Decalogue
closed
[DOC] Features for easier trading
documentation uncorrelated
**Type of Proposal** - [&check;] Feature Request (e.g. functionality) - [ ] Economic Model - [ ] Underlying Technology (e.g. performance) - [ ] Application Development - [ ] Others **Background** Ease of use for traders, especially those who are new to KuCoin, based on my personal experience when I first using Kucoin. **Describe The Proposal (No more than 50 words)** Simplifying futures funds transfer and auto deposit margin. **Business Model (Optional for underlying technology)** Making transferring funds and futures trading easier for traders so they will feel more comfortable in using KuCoin and hopefully will stay and be a loyal customer. **Implementation Steps (Optional for economic model)** 1. Funds transfer As of now, traders have to transfer their funds from main account to trading account first then transfer it again to futures account in order to trade (that was what I experienced and got confused at first because I thought my funds went missing). It would be more convenient for traders if they can directly transfer their funds from main account to trading, futures, or margin account, so they only have to do it one time instead of two times. 2. Auto deposit margin In futures trading, traders can enable "auto deposit margin" feature after their order got filled (again, this was based on personal experience). It would be a hassle if a trader isn't online at the time their order got filled, resulting they can't enable the "auto deposit margin" feature. Having that feature available during order placing will be a neat solution, so traders won't have to wait until their order got filled to activate it. **Expected Outcome** Traders will have a smoother, hassle free trade while trading on KuCoin. **Timetable (Optional)** As soon as possible, preferably once KuChain goes live. **Additional Context (Optional)** N/A **Contact Information** ridhataqobalallahlah@gmail.com
1.0
[DOC] Features for easier trading - **Type of Proposal** - [&check;] Feature Request (e.g. functionality) - [ ] Economic Model - [ ] Underlying Technology (e.g. performance) - [ ] Application Development - [ ] Others **Background** Ease of use for traders, especially those who are new to KuCoin, based on my personal experience when I first using Kucoin. **Describe The Proposal (No more than 50 words)** Simplifying futures funds transfer and auto deposit margin. **Business Model (Optional for underlying technology)** Making transferring funds and futures trading easier for traders so they will feel more comfortable in using KuCoin and hopefully will stay and be a loyal customer. **Implementation Steps (Optional for economic model)** 1. Funds transfer As of now, traders have to transfer their funds from main account to trading account first then transfer it again to futures account in order to trade (that was what I experienced and got confused at first because I thought my funds went missing). It would be more convenient for traders if they can directly transfer their funds from main account to trading, futures, or margin account, so they only have to do it one time instead of two times. 2. Auto deposit margin In futures trading, traders can enable "auto deposit margin" feature after their order got filled (again, this was based on personal experience). It would be a hassle if a trader isn't online at the time their order got filled, resulting they can't enable the "auto deposit margin" feature. Having that feature available during order placing will be a neat solution, so traders won't have to wait until their order got filled to activate it. **Expected Outcome** Traders will have a smoother, hassle free trade while trading on KuCoin. **Timetable (Optional)** As soon as possible, preferably once KuChain goes live. **Additional Context (Optional)** N/A **Contact Information** ridhataqobalallahlah@gmail.com
non_process
features for easier trading type of proposal feature request e g functionality economic model underlying technology e g performance application development others background ease of use for traders especially those who are new to kucoin based on my personal experience when i first using kucoin describe the proposal no more than words simplifying futures funds transfer and auto deposit margin business model optional for underlying technology making transferring funds and futures trading easier for traders so they will feel more comfortable in using kucoin and hopefully will stay and be a loyal customer implementation steps optional for economic model funds transfer as of now traders have to transfer their funds from main account to trading account first then transfer it again to futures account in order to trade that was what i experienced and got confused at first because i thought my funds went missing it would be more convenient for traders if they can directly transfer their funds from main account to trading futures or margin account so they only have to do it one time instead of two times auto deposit margin in futures trading traders can enable auto deposit margin feature after their order got filled again this was based on personal experience it would be a hassle if a trader isn t online at the time their order got filled resulting they can t enable the auto deposit margin feature having that feature available during order placing will be a neat solution so traders won t have to wait until their order got filled to activate it expected outcome traders will have a smoother hassle free trade while trading on kucoin timetable optional as soon as possible preferably once kuchain goes live additional context optional n a contact information ridhataqobalallahlah gmail com
0
7,912
11,092,325,985
IssuesEvent
2019-12-15 18:14:27
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Add new `sync_response` processor
annoying enhancement processors
It would be useful to have a `sync_response` processor that would enable more customized behaviour for pipelines containing responses. This is going to be simple to implement but we need to add extra care that it won't introduce race conditions (post-buffer alterations).
1.0
Add new `sync_response` processor - It would be useful to have a `sync_response` processor that would enable more customized behaviour for pipelines containing responses. This is going to be simple to implement but we need to add extra care that it won't introduce race conditions (post-buffer alterations).
process
add new sync response processor it would be useful to have a sync response processor that would enable more customized behaviour for pipelines containing responses this is going to be simple to implement but we need to add extra care that it won t introduce race conditions post buffer alterations
1
29,800
2,717,434,445
IssuesEvent
2015-04-11 08:29:55
codenameone/CodenameOne
https://api.github.com/repos/codenameone/CodenameOne
closed
RFE: line seperator component
Priority-Medium Type-Enhancement
Original [issue 264](https://code.google.com/p/codenameone/issues/detail?id=264) created by codenameone on 2012-07-15T22:33:51.000Z: <b>What steps will reproduce the problem?</b> 1. (Line) Seperator Component in Designer <b>What is the expected output? What do you see instead?</b> I would like to have a Seperator Component in the Designer that can seperate for example a list of commands through a line. I think it's good for the usability, because you can make interfaces more clearly. In the Designer the iPhone preview of the ComponentGroup does this, but on the simulator i cannot see these grey lines. Would be nice to be able to make them myself... <b>What version of the product are you using? On what operating system?</b> latest cn1, w7 <b>Please provide any additional information below.</b> -
1.0
RFE: line seperator component - Original [issue 264](https://code.google.com/p/codenameone/issues/detail?id=264) created by codenameone on 2012-07-15T22:33:51.000Z: <b>What steps will reproduce the problem?</b> 1. (Line) Seperator Component in Designer <b>What is the expected output? What do you see instead?</b> I would like to have a Seperator Component in the Designer that can seperate for example a list of commands through a line. I think it's good for the usability, because you can make interfaces more clearly. In the Designer the iPhone preview of the ComponentGroup does this, but on the simulator i cannot see these grey lines. Would be nice to be able to make them myself... <b>What version of the product are you using? On what operating system?</b> latest cn1, w7 <b>Please provide any additional information below.</b> -
non_process
rfe line seperator component original created by codenameone on what steps will reproduce the problem line seperator component in designer what is the expected output what do you see instead i would like to have a seperator component in the designer that can seperate for example a list of commands through a line i think it s good for the usability because you can make interfaces more clearly in the designer the iphone preview of the componentgroup does this but on the simulator i cannot see these grey lines would be nice to be able to make them myself what version of the product are you using on what operating system latest please provide any additional information below
0
57,908
16,136,922,287
IssuesEvent
2021-04-29 13:01:38
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
Tree Selection does not fulfil all accessibility criteria (WCAG)
defect
**Issue Description** Primefaces' Tree Selection component does nut fulfil the WCAG accessibility criterion [4.1.1](https://www.w3.org/WAI/WCAG21/Understanding/parsing.html) as it renders syntactically incorrect HTML code. The problems seems to be that in complex trees which include components like checkboxes the included components are rendered as 'div' elements while the tree itself as well as its tree nodes are rendered as 'span' elements ('div' elements are not allowed as child elements of 'span' elements). **To Reproduce** Steps to reproduce the behavior: 1. Open showcase [Tree Selection](https://www.primefaces.org/showcase/ui/data/tree/selection.xhtml?jfwid=4c890). 2. Use bookmarklet "Check serialized DOM of current page" from https://validator.w3.org/nu/about.html#extras to run Nu Html Checker on the showcase. 3. Drag and drop bookmarklet "Check for WCAG 2.0 parsing compliance" from the same page to your browser's bookmark toolbar and execute it on the website of Nu Html Checker which you opened in Step 2. This bookmarklet will filter the results of Nu Html Checker so that only WCAG relevant results remain. 4. After running both bookmarklets, if the Nu Html Checker still shows error messages, the syntax of the checked HTML code is not correct. **Expected behavior** To fulfil the accessibility criterion Nu Html Checker should show no error messages after checking the tree's HTML code. **Example XHTML** `<p:tree value="#{treeSelectionView.root3}" var="doc" selectionMode="checkbox" selection="#{treeSelectionView.selectedNodes2}"> <p:treeNode expandedIcon="pi pi-folder-open" collapsedIcon="pi pi-folder"> <h:outputText value="#{doc.name}"/> </p:treeNode> </p:tree>`
1.0
Tree Selection does not fulfil all accessibility criteria (WCAG) - **Issue Description** Primefaces' Tree Selection component does nut fulfil the WCAG accessibility criterion [4.1.1](https://www.w3.org/WAI/WCAG21/Understanding/parsing.html) as it renders syntactically incorrect HTML code. The problems seems to be that in complex trees which include components like checkboxes the included components are rendered as 'div' elements while the tree itself as well as its tree nodes are rendered as 'span' elements ('div' elements are not allowed as child elements of 'span' elements). **To Reproduce** Steps to reproduce the behavior: 1. Open showcase [Tree Selection](https://www.primefaces.org/showcase/ui/data/tree/selection.xhtml?jfwid=4c890). 2. Use bookmarklet "Check serialized DOM of current page" from https://validator.w3.org/nu/about.html#extras to run Nu Html Checker on the showcase. 3. Drag and drop bookmarklet "Check for WCAG 2.0 parsing compliance" from the same page to your browser's bookmark toolbar and execute it on the website of Nu Html Checker which you opened in Step 2. This bookmarklet will filter the results of Nu Html Checker so that only WCAG relevant results remain. 4. After running both bookmarklets, if the Nu Html Checker still shows error messages, the syntax of the checked HTML code is not correct. **Expected behavior** To fulfil the accessibility criterion Nu Html Checker should show no error messages after checking the tree's HTML code. **Example XHTML** `<p:tree value="#{treeSelectionView.root3}" var="doc" selectionMode="checkbox" selection="#{treeSelectionView.selectedNodes2}"> <p:treeNode expandedIcon="pi pi-folder-open" collapsedIcon="pi pi-folder"> <h:outputText value="#{doc.name}"/> </p:treeNode> </p:tree>`
non_process
tree selection does not fulfil all accessibility criteria wcag issue description primefaces tree selection component does nut fulfil the wcag accessibility criterion as it renders syntactically incorrect html code the problems seems to be that in complex trees which include components like checkboxes the included components are rendered as div elements while the tree itself as well as its tree nodes are rendered as span elements div elements are not allowed as child elements of span elements to reproduce steps to reproduce the behavior open showcase use bookmarklet check serialized dom of current page from to run nu html checker on the showcase drag and drop bookmarklet check for wcag parsing compliance from the same page to your browser s bookmark toolbar and execute it on the website of nu html checker which you opened in step this bookmarklet will filter the results of nu html checker so that only wcag relevant results remain after running both bookmarklets if the nu html checker still shows error messages the syntax of the checked html code is not correct expected behavior to fulfil the accessibility criterion nu html checker should show no error messages after checking the tree s html code example xhtml p tree value treeselectionview var doc selectionmode checkbox selection treeselectionview
0
12,816
15,190,289,263
IssuesEvent
2021-02-15 17:41:23
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
closed
If user deletes the IAM role, deleting an s3 source fails
bug p0 team:data processing
### Describe the bug A clear and concise description of what the bug is. ### Steps to reproduce 1. Onboard an s3 source, enable Panther to manage bucket notifications 2. Finish source onboarding 3. Delete the IAM role 4. Delete source - Removing bucket notifications will fail because the role is absent and source deletion will fail as well ### Expected behavior Deleting a source should succeed even if the IAM role is deleted ### Environment How are you deploying or using Panther? - Panther version or commit: master, 1.16.0-dev
1.0
If user deletes the IAM role, deleting an s3 source fails - ### Describe the bug A clear and concise description of what the bug is. ### Steps to reproduce 1. Onboard an s3 source, enable Panther to manage bucket notifications 2. Finish source onboarding 3. Delete the IAM role 4. Delete source - Removing bucket notifications will fail because the role is absent and source deletion will fail as well ### Expected behavior Deleting a source should succeed even if the IAM role is deleted ### Environment How are you deploying or using Panther? - Panther version or commit: master, 1.16.0-dev
process
if user deletes the iam role deleting an source fails describe the bug a clear and concise description of what the bug is steps to reproduce onboard an source enable panther to manage bucket notifications finish source onboarding delete the iam role delete source removing bucket notifications will fail because the role is absent and source deletion will fail as well expected behavior deleting a source should succeed even if the iam role is deleted environment how are you deploying or using panther panther version or commit master dev
1
508,291
14,697,981,439
IssuesEvent
2021-01-04 05:14:35
boomerang-io/roadmap
https://api.github.com/repos/boomerang-io/roadmap
opened
No network flag security context for Custom Task or All Tasks
enhancement priority: high revision needed
**Is your request related to a problem? Please describe.** The ability to set low level kubernetes security contexts such as the No Network flag. **Describe the solution you'd like** Be able to define specific low level security contexts at either the Custom Task level or the System wide level **Describe the benefits or justification for this request** This will allow authors to create workflows that may use containers that are created by external parties and are not fully trusted **Additional context** Will need to determine how this is implemented and at which level. Right now security contexts are hardcoded in the controller. FYI: @simon-harris-ibm
1.0
No network flag security context for Custom Task or All Tasks - **Is your request related to a problem? Please describe.** The ability to set low level kubernetes security contexts such as the No Network flag. **Describe the solution you'd like** Be able to define specific low level security contexts at either the Custom Task level or the System wide level **Describe the benefits or justification for this request** This will allow authors to create workflows that may use containers that are created by external parties and are not fully trusted **Additional context** Will need to determine how this is implemented and at which level. Right now security contexts are hardcoded in the controller. FYI: @simon-harris-ibm
non_process
no network flag security context for custom task or all tasks is your request related to a problem please describe the ability to set low level kubernetes security contexts such as the no network flag describe the solution you d like be able to define specific low level security contexts at either the custom task level or the system wide level describe the benefits or justification for this request this will allow authors to create workflows that may use containers that are created by external parties and are not fully trusted additional context will need to determine how this is implemented and at which level right now security contexts are hardcoded in the controller fyi simon harris ibm
0
104,863
4,226,242,267
IssuesEvent
2016-07-02 10:03:11
FAC-GM/app
https://api.github.com/repos/FAC-GM/app
closed
Re-designing the home page + login page
priority-5 UI
Re-designing the home page + login page according to PO wireframes.
1.0
Re-designing the home page + login page - Re-designing the home page + login page according to PO wireframes.
non_process
re designing the home page login page re designing the home page login page according to po wireframes
0
11,027
13,822,892,707
IssuesEvent
2020-10-13 06:06:15
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
20.09 Release notes
0.kind: enhancement 6.topic: release process
This thread is for any release-worthy notes which may have not have made their way into https://github.com/NixOS/nixpkgs/blob/master/nixos/doc/manual/release-notes/rl-2009.xml yet. Please leave a summary and any relevant links to the items. I will try and go through them before the release to ensure the notes are in order. List to make items easier to track: - [x] Python35 was removed - [x] Agda changes - [x] Amdvlk can be added to `hardware.opengl.extraPackages` - [x] Cinnamon desktop was added - [x] Fontconfig was bumped, and reworked on NixOS - [x] Nvidia Optimus/Prime is fully usable now
1.0
20.09 Release notes - This thread is for any release-worthy notes which may have not have made their way into https://github.com/NixOS/nixpkgs/blob/master/nixos/doc/manual/release-notes/rl-2009.xml yet. Please leave a summary and any relevant links to the items. I will try and go through them before the release to ensure the notes are in order. List to make items easier to track: - [x] Python35 was removed - [x] Agda changes - [x] Amdvlk can be added to `hardware.opengl.extraPackages` - [x] Cinnamon desktop was added - [x] Fontconfig was bumped, and reworked on NixOS - [x] Nvidia Optimus/Prime is fully usable now
process
release notes this thread is for any release worthy notes which may have not have made their way into yet please leave a summary and any relevant links to the items i will try and go through them before the release to ensure the notes are in order list to make items easier to track was removed agda changes amdvlk can be added to hardware opengl extrapackages cinnamon desktop was added fontconfig was bumped and reworked on nixos nvidia optimus prime is fully usable now
1
954
3,418,291,273
IssuesEvent
2015-12-08 01:05:40
wekan/wekan
https://api.github.com/repos/wekan/wekan
opened
Release v0.10
Meta:Release-process
Hi, I've released the [first release candidate](https://github.com/wekan/wekan/releases/tag/v0.10.0-rc1) of Wekan v0.10. The release link contains the Sandstorm `spk`, the Nodejs `tar.gz` application, and the Docker image. You can [read the release notes here](https://github.com/wekan/wekan/blob/master/History.md#v010). This release includes some contributions by @AlexanderS, @fisle, @floatinghotpot, @FuzzyWuzzie, @mnutt, @ndarilek, @SirCmpwn, and @xavierpriour, thank you all very much. Issues remaining before the final v0.10 are tracked using the corresponding [GitHub milestone](https://github.com/wekan/wekan/milestones/Release%200.10). We need your help to test the application, report bugs, and translate the new features to your local languages! If you don’t want to install Wekan on your own server, I’ll deploy an instance on https://wekan.io sometimes in the coming days.
1.0
Release v0.10 - Hi, I've released the [first release candidate](https://github.com/wekan/wekan/releases/tag/v0.10.0-rc1) of Wekan v0.10. The release link contains the Sandstorm `spk`, the Nodejs `tar.gz` application, and the Docker image. You can [read the release notes here](https://github.com/wekan/wekan/blob/master/History.md#v010). This release includes some contributions by @AlexanderS, @fisle, @floatinghotpot, @FuzzyWuzzie, @mnutt, @ndarilek, @SirCmpwn, and @xavierpriour, thank you all very much. Issues remaining before the final v0.10 are tracked using the corresponding [GitHub milestone](https://github.com/wekan/wekan/milestones/Release%200.10). We need your help to test the application, report bugs, and translate the new features to your local languages! If you don’t want to install Wekan on your own server, I’ll deploy an instance on https://wekan.io sometimes in the coming days.
process
release hi i ve released the of wekan the release link contains the sandstorm spk the nodejs tar gz application and the docker image you can this release includes some contributions by alexanders fisle floatinghotpot fuzzywuzzie mnutt ndarilek sircmpwn and xavierpriour thank you all very much issues remaining before the final are tracked using the corresponding we need your help to test the application report bugs and translate the new features to your local languages if you don’t want to install wekan on your own server i’ll deploy an instance on sometimes in the coming days
1
12,729
15,099,298,986
IssuesEvent
2021-02-08 02:07:53
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Allowing drop='first' and handle_unknown='ignore' in OneHotEncoding
module:preprocessing
When using OneHotEncoding it is a common feature to drop a column in order to prevent multi-collinearity. This is done with the drop='first' flag. On the other hand, when cross validating a model, it can happen that not every category appears on train split. In this setting, it is usual to just assign 0 to all columns. This is done with handle_unknown='ignore'. Both flags cannot be set as shows as it leads to: ``` ValueError: `handle_unknown` must be 'error' when the drop parameter is specified, as both would create categories that are all zero. ``` It is true that using drop='first' and handle_unknown='ignore' leads to some degree of ambiguity. But I don't see any other way to deal with multi-collinearity and unknown classes at training at the same time. I believe the current behavior is inadequate. It would be better to enable setting both flags at the same time. [This post](https://geoffruddock.com/one-hot-encoding-plus-linear-regression-equals-multi-collinearity/#but-it-doesnt-play-nicely-with-cv-pipelines) and [this question](https://stackoverflow.com/q/60008477/7502752) have found the same thing problematic.
1.0
Allowing drop='first' and handle_unknown='ignore' in OneHotEncoding - When using OneHotEncoding it is a common feature to drop a column in order to prevent multi-collinearity. This is done with the drop='first' flag. On the other hand, when cross validating a model, it can happen that not every category appears on train split. In this setting, it is usual to just assign 0 to all columns. This is done with handle_unknown='ignore'. Both flags cannot be set as shows as it leads to: ``` ValueError: `handle_unknown` must be 'error' when the drop parameter is specified, as both would create categories that are all zero. ``` It is true that using drop='first' and handle_unknown='ignore' leads to some degree of ambiguity. But I don't see any other way to deal with multi-collinearity and unknown classes at training at the same time. I believe the current behavior is inadequate. It would be better to enable setting both flags at the same time. [This post](https://geoffruddock.com/one-hot-encoding-plus-linear-regression-equals-multi-collinearity/#but-it-doesnt-play-nicely-with-cv-pipelines) and [this question](https://stackoverflow.com/q/60008477/7502752) have found the same thing problematic.
process
allowing drop first and handle unknown ignore in onehotencoding when using onehotencoding it is a common feature to drop a column in order to prevent multi collinearity this is done with the drop first flag on the other hand when cross validating a model it can happen that not every category appears on train split in this setting it is usual to just assign to all columns this is done with handle unknown ignore both flags cannot be set as shows as it leads to valueerror handle unknown must be error when the drop parameter is specified as both would create categories that are all zero it is true that using drop first and handle unknown ignore leads to some degree of ambiguity but i don t see any other way to deal with multi collinearity and unknown classes at training at the same time i believe the current behavior is inadequate it would be better to enable setting both flags at the same time and have found the same thing problematic
1
84,981
3,683,069,368
IssuesEvent
2016-02-24 12:32:31
wp-property/wp-property
https://api.github.com/repos/wp-property/wp-property
opened
Can't delete lats item in Meta and Terms section
priority/high type/bug
There should be at least one. User notices from wordpress - https://wordpress.org/support/topic/cant-properly-remove-community-features-terms?replies=4#post-8070870
1.0
Can't delete lats item in Meta and Terms section - There should be at least one. User notices from wordpress - https://wordpress.org/support/topic/cant-properly-remove-community-features-terms?replies=4#post-8070870
non_process
can t delete lats item in meta and terms section there should be at least one user notices from wordpress
0
100,320
12,514,843,239
IssuesEvent
2020-06-03 06:24:37
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Hide yellow banner
area/dashboard kind/user-story solution/studio/designer
## Description Hide the yellow stuff on dashboard since the content is no longer relevant. ## Screenshots ![image](https://user-images.githubusercontent.com/6088624/82043084-44703600-96ab-11ea-87d7-964a9c4812b1.png) ## Considerations Only comment out so it's easy to put back. ## Acceptance criteria Yellow banner is gone. ## Tasks - [ ] Remove banner (comment out) - [ ] Test
1.0
Hide yellow banner - ## Description Hide the yellow stuff on dashboard since the content is no longer relevant. ## Screenshots ![image](https://user-images.githubusercontent.com/6088624/82043084-44703600-96ab-11ea-87d7-964a9c4812b1.png) ## Considerations Only comment out so it's easy to put back. ## Acceptance criteria Yellow banner is gone. ## Tasks - [ ] Remove banner (comment out) - [ ] Test
non_process
hide yellow banner description hide the yellow stuff on dashboard since the content is no longer relevant screenshots considerations only comment out so it s easy to put back acceptance criteria yellow banner is gone tasks remove banner comment out test
0
23,804
3,870,942,345
IssuesEvent
2016-04-11 07:36:40
geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE
https://api.github.com/repos/geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE
closed
a45kw6vEeeNoXIqYQ5fZIz2ddDWg8Fu3Q/+yx96hgmve51lCYgVUw3DGJEDqhgYRKT2yNP5qAzdKY+FOApdtsgcXjZs6pCF86t3nDfpDORFAf2ou+TQ/Qhw8M9aOL2xy6CV6OG3YoO2YBef6MXt59ZMS8UO8uxt00syItffwpNo=
design
EThWldcBapO7q/nWLiBxYLEPPp+7IglaA3AugKF557F2XQgSL7HlAKiZE6FGSRYfkYLuM9cde1RUA++Y9+ciA6dL4Qs9J/cUTvozU+JnVdJv18cJ4QZYK1tSo8lLAVM0YkCFjJ4Tm5fXmQS87OZodfLNWQ5u6aRksAgg/v67u0FTH0rvTFckM/OEDyg8ZuSYSTHZHvQo3M+K77qdG+1dbmZEJId+1uAIQH5CLmxQTt3OBTaIPCPEJF9U4yBEZEkMwpn9oUrxCEajy+Al0+ZwISJlX3LBYab/YClVa+HiJKjJyeIH3Wwy/K37b6aVTIZLC6aFg/BF4P3ISpOAuVT0eLEi6dAoRWNI+0AzXwklYLo9E9hqLhkV2cSNwU6LjIYH0vvo/kJK4sscai18nFIf9z8ve0CX4bkNvoouYlmLFGyPuhSB4CU27GmHjQZ78LM9Yzvlho5VbJlbV1hjSp1yl7FZb305+GP5F3/1hEt1cbE+Gjrv3EkrOB/WYNH44RV/81oUvCv0pEmmMf1D/UiLZcUaV1ZVE2DYN5aV+p+HspA=
1.0
a45kw6vEeeNoXIqYQ5fZIz2ddDWg8Fu3Q/+yx96hgmve51lCYgVUw3DGJEDqhgYRKT2yNP5qAzdKY+FOApdtsgcXjZs6pCF86t3nDfpDORFAf2ou+TQ/Qhw8M9aOL2xy6CV6OG3YoO2YBef6MXt59ZMS8UO8uxt00syItffwpNo= - EThWldcBapO7q/nWLiBxYLEPPp+7IglaA3AugKF557F2XQgSL7HlAKiZE6FGSRYfkYLuM9cde1RUA++Y9+ciA6dL4Qs9J/cUTvozU+JnVdJv18cJ4QZYK1tSo8lLAVM0YkCFjJ4Tm5fXmQS87OZodfLNWQ5u6aRksAgg/v67u0FTH0rvTFckM/OEDyg8ZuSYSTHZHvQo3M+K77qdG+1dbmZEJId+1uAIQH5CLmxQTt3OBTaIPCPEJF9U4yBEZEkMwpn9oUrxCEajy+Al0+ZwISJlX3LBYab/YClVa+HiJKjJyeIH3Wwy/K37b6aVTIZLC6aFg/BF4P3ISpOAuVT0eLEi6dAoRWNI+0AzXwklYLo9E9hqLhkV2cSNwU6LjIYH0vvo/kJK4sscai18nFIf9z8ve0CX4bkNvoouYlmLFGyPuhSB4CU27GmHjQZ78LM9Yzvlho5VbJlbV1hjSp1yl7FZb305+GP5F3/1hEt1cbE+Gjrv3EkrOB/WYNH44RV/81oUvCv0pEmmMf1D/UiLZcUaV1ZVE2DYN5aV+p+HspA=
non_process
tq nwlibxyleppp cutvozu yclva p hspa
0
109,331
9,378,327,358
IssuesEvent
2019-04-04 12:38:58
kcigeospatial/Fred_Co_Land-Management
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
closed
Planning-verification letters
Ready For Retest
verification letters should not go to finalized before the condition requiring letter to be sent is approved. ME
1.0
Planning-verification letters - verification letters should not go to finalized before the condition requiring letter to be sent is approved. ME
non_process
planning verification letters verification letters should not go to finalized before the condition requiring letter to be sent is approved me
0
72,528
3,386,806,491
IssuesEvent
2015-11-27 21:33:23
arcualberta/TAPoR
https://api.github.com/repos/arcualberta/TAPoR
closed
New tags not returning tools
high priority
After adding new tags, the tag doesn't return the tool as a search result when clicked as a link or when searched from the search bar.
1.0
New tags not returning tools - After adding new tags, the tag doesn't return the tool as a search result when clicked as a link or when searched from the search bar.
non_process
new tags not returning tools after adding new tags the tag doesn t return the tool as a search result when clicked as a link or when searched from the search bar
0
15,841
20,028,187,313
IssuesEvent
2022-02-02 00:26:45
googleapis/java-translate
https://api.github.com/repos/googleapis/java-translate
closed
com.example.translate.BatchTranslateTextWithGlossaryTests: testBatchTranslateTextWithGlossary failed
priority: p2 type: process api: translate flakybot: issue flakybot: flaky
Note: #567 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b7ad2372edd50a38d703efb8305d7c5e71145409 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7385ee88-4195-4405-a11e-8d439d404617), [Sponge](http://sponge2/7385ee88-4195-4405-a11e-8d439d404617) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:93) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateTextWithGlossary.batchTranslateTextWithGlossary(BatchTranslateTextWithGlossary.java:110) at com.example.translate.BatchTranslateTextWithGlossaryTests.testBatchTranslateTextWithGlossary(BatchTranslateTextWithGlossaryTests.java:123) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:553) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:68) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:739) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:718) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 17 more </pre></details>
1.0
com.example.translate.BatchTranslateTextWithGlossaryTests: testBatchTranslateTextWithGlossary failed - Note: #567 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b7ad2372edd50a38d703efb8305d7c5e71145409 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7385ee88-4195-4405-a11e-8d439d404617), [Sponge](http://sponge2/7385ee88-4195-4405-a11e-8d439d404617) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:93) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateTextWithGlossary.batchTranslateTextWithGlossary(BatchTranslateTextWithGlossary.java:110) at com.example.translate.BatchTranslateTextWithGlossaryTests.testBatchTranslateTextWithGlossary(BatchTranslateTextWithGlossaryTests.java:123) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:553) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:68) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:739) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:718) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_GLOSSARY_OUTPUT_6c3a18ff-5bce-4769-88f7-88ca31d2995c/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 17 more </pre></details>
process
com example translate batchtranslatetextwithglossarytests testbatchtranslatetextwithglossary failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output java util concurrent executionexception com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation glossary output at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com example translate batchtranslatetextwithglossary batchtranslatetextwithglossary batchtranslatetextwithglossary java at com example translate batchtranslatetextwithglossarytests testbatchtranslatetextwithglossary batchtranslatetextwithglossarytests java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at com google cloud testing multipleattemptsrule evaluate multipleattemptsrule java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation glossary output at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation glossary output at io grpc status asruntimeexception status java more
1
332,191
24,338,204,765
IssuesEvent
2022-10-01 10:57:21
Crequency/KitX
https://api.github.com/repos/Crequency/KitX
closed
关于插件的实现
documentation
按我的理解是插件编译为托管dll,然后调用预先规定名称的函数吗 似乎是直接用GetExportedValues导入为IIdentityInterface对象后启动controller然后向dashboard发送json化的插件信息? 也就是说每一个插件都有一个loader相对应吗 这样的话有些无法编译为托管dll的语言是怎么支持的?插件间的交互如自定义函数按名调用是怎么实现的? (我不太熟悉csharp,没有在源码里找到相应的功能)
1.0
关于插件的实现 - 按我的理解是插件编译为托管dll,然后调用预先规定名称的函数吗 似乎是直接用GetExportedValues导入为IIdentityInterface对象后启动controller然后向dashboard发送json化的插件信息? 也就是说每一个插件都有一个loader相对应吗 这样的话有些无法编译为托管dll的语言是怎么支持的?插件间的交互如自定义函数按名调用是怎么实现的? (我不太熟悉csharp,没有在源码里找到相应的功能)
non_process
关于插件的实现 按我的理解是插件编译为托管dll,然后调用预先规定名称的函数吗 似乎是直接用getexportedvalues导入为iidentityinterface对象后启动controller然后向dashboard发送json化的插件信息? 也就是说每一个插件都有一个loader相对应吗 这样的话有些无法编译为托管dll的语言是怎么支持的?插件间的交互如自定义函数按名调用是怎么实现的? (我不太熟悉csharp,没有在源码里找到相应的功能)
0
195,436
22,339,631,699
IssuesEvent
2022-06-14 22:33:24
vincenzodistasio97/events-manager-io
https://api.github.com/repos/vincenzodistasio97/events-manager-io
closed
CVE-2019-10752 (High) detected in sequelize-4.37.6.tgz - autoclosed
security vulnerability
## CVE-2019-10752 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sequelize-4.37.6.tgz</b></p></summary> <p>Multi dialect ORM for Node.JS</p> <p>Library home page: <a href="https://registry.npmjs.org/sequelize/-/sequelize-4.37.6.tgz">https://registry.npmjs.org/sequelize/-/sequelize-4.37.6.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sequelize/package.json</p> <p> Dependency Hierarchy: - :x: **sequelize-4.37.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/events-manager-io/commit/34b4ee7777ec330308085b59cefc667c68e51123">34b4ee7777ec330308085b59cefc667c68e51123</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Sequelize, all versions prior to version 4.44.3 and 5.15.1, is vulnerable to SQL Injection due to sequelize.json() helper function not escaping values properly when formatting sub paths for JSON queries for MySQL, MariaDB and SQLite. <p>Publish Date: 2019-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10752>CVE-2019-10752</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10752">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10752</a></p> <p>Release Date: 2019-10-21</p> <p>Fix Resolution: 4.44.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10752 (High) detected in sequelize-4.37.6.tgz - autoclosed - ## CVE-2019-10752 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sequelize-4.37.6.tgz</b></p></summary> <p>Multi dialect ORM for Node.JS</p> <p>Library home page: <a href="https://registry.npmjs.org/sequelize/-/sequelize-4.37.6.tgz">https://registry.npmjs.org/sequelize/-/sequelize-4.37.6.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sequelize/package.json</p> <p> Dependency Hierarchy: - :x: **sequelize-4.37.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/events-manager-io/commit/34b4ee7777ec330308085b59cefc667c68e51123">34b4ee7777ec330308085b59cefc667c68e51123</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Sequelize, all versions prior to version 4.44.3 and 5.15.1, is vulnerable to SQL Injection due to sequelize.json() helper function not escaping values properly when formatting sub paths for JSON queries for MySQL, MariaDB and SQLite. <p>Publish Date: 2019-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10752>CVE-2019-10752</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10752">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10752</a></p> <p>Release Date: 2019-10-21</p> <p>Fix Resolution: 4.44.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in sequelize tgz autoclosed cve high severity vulnerability vulnerable library sequelize tgz multi dialect orm for node js library home page a href path to dependency file package json path to vulnerable library node modules sequelize package json dependency hierarchy x sequelize tgz vulnerable library found in head commit a href found in base branch dev vulnerability details sequelize all versions prior to version and is vulnerable to sql injection due to sequelize json helper function not escaping values properly when formatting sub paths for json queries for mysql mariadb and sqlite publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
19,799
3,497,607,435
IssuesEvent
2016-01-06 02:24:51
meren/anvio
https://api.github.com/repos/meren/anvio
opened
Switching to cherrypy: user management re-design.
design enhancement server
OK. This is a big change. We have been having some random issues with the server performance as we were using the default backend from within bottle (`wsgiref`; the single-threaded backend). I think in the long run it is much better to rely on a multithreaded backend like cherrypy (see the reasoning [here](http://bottlepy.org/docs/dev/deployment.html#switching-the-server-backend)), since the default is only good for testing and development. So we have to do this sooner or later. There are three places in the codebase where the bottle server is called: * anvi-interactive * anvi-refine * anvi-server I switched to cherrypy in all of them by simply making this change (which will require a `sudo pip install cherrypy` for it to work): ``` diff - server_process = Process(target=run_server, kwargs={'host': ip, 'port': port, 'quiet': True}) + server_process = Process(target=run_server, kwargs={'host': ip, 'port': port, 'quiet': True, 'server': 'cherrypy'}) ``` Due to the database design, anvi-interactive and anvi-refine worked without any issues with the new multi-threaded backend. However, anvi-server started having some issues. When I looked into these things, I realized that the UserMGMT class initiates a database connection, and uses it for everything. Because we are relying on self-contained database files and not a database server, this is prone to a lot of issues going forward. I changed the design a bit, and created a `UsersDB` class in the usermanagement.py. Now UserMGMT class gets a copy of this one, and all database calls go through that class. I started changing things just to test in UserMGMT, but haven't really finished. I will commit my changes so far in a second, but following examples the way things should look like now. #### fetchone ``` diff - p = (token, ) - response = self.cursor.execute("SELECT * FROM users WHERE token=?", p) - user = response.fetchone() + user = self.users_db.fetchone("SELECT * FROM users WHERE token=?", (token, )) ``` #### fetchall ``` diff - p = (user['login'], ) - response = self.cursor.execute("SELECT name FROM projects WHERE user=?", p) - projects = response.fetchall() + projects = self.users_db.fetchall("SELECT name FROM projects WHERE user=?", (user['login'], )) ``` #### execute ``` diff - p = (token, user['login'], ) - self.cursor.execute("UPDATE users SET token=? WHERE login=?", p) - self.conn.commit() + self.users_db.execute("UPDATE users SET token=? WHERE login=?", (token, user['login'], )) ``` I hope this is agreeable. Best,
1.0
Switching to cherrypy: user management re-design. - OK. This is a big change. We have been having some random issues with the server performance as we were using the default backend from within bottle (`wsgiref`; the single-threaded backend). I think in the long run it is much better to rely on a multithreaded backend like cherrypy (see the reasoning [here](http://bottlepy.org/docs/dev/deployment.html#switching-the-server-backend)), since the default is only good for testing and development. So we have to do this sooner or later. There are three places in the codebase where the bottle server is called: * anvi-interactive * anvi-refine * anvi-server I switched to cherrypy in all of them by simply making this change (which will require a `sudo pip install cherrypy` for it to work): ``` diff - server_process = Process(target=run_server, kwargs={'host': ip, 'port': port, 'quiet': True}) + server_process = Process(target=run_server, kwargs={'host': ip, 'port': port, 'quiet': True, 'server': 'cherrypy'}) ``` Due to the database design, anvi-interactive and anvi-refine worked without any issues with the new multi-threaded backend. However, anvi-server started having some issues. When I looked into these things, I realized that the UserMGMT class initiates a database connection, and uses it for everything. Because we are relying on self-contained database files and not a database server, this is prone to a lot of issues going forward. I changed the design a bit, and created a `UsersDB` class in the usermanagement.py. Now UserMGMT class gets a copy of this one, and all database calls go through that class. I started changing things just to test in UserMGMT, but haven't really finished. I will commit my changes so far in a second, but following examples the way things should look like now. #### fetchone ``` diff - p = (token, ) - response = self.cursor.execute("SELECT * FROM users WHERE token=?", p) - user = response.fetchone() + user = self.users_db.fetchone("SELECT * FROM users WHERE token=?", (token, )) ``` #### fetchall ``` diff - p = (user['login'], ) - response = self.cursor.execute("SELECT name FROM projects WHERE user=?", p) - projects = response.fetchall() + projects = self.users_db.fetchall("SELECT name FROM projects WHERE user=?", (user['login'], )) ``` #### execute ``` diff - p = (token, user['login'], ) - self.cursor.execute("UPDATE users SET token=? WHERE login=?", p) - self.conn.commit() + self.users_db.execute("UPDATE users SET token=? WHERE login=?", (token, user['login'], )) ``` I hope this is agreeable. Best,
non_process
switching to cherrypy user management re design ok this is a big change we have been having some random issues with the server performance as we were using the default backend from within bottle wsgiref the single threaded backend i think in the long run it is much better to rely on a multithreaded backend like cherrypy see the reasoning since the default is only good for testing and development so we have to do this sooner or later there are three places in the codebase where the bottle server is called anvi interactive anvi refine anvi server i switched to cherrypy in all of them by simply making this change which will require a sudo pip install cherrypy for it to work diff server process process target run server kwargs host ip port port quiet true server process process target run server kwargs host ip port port quiet true server cherrypy due to the database design anvi interactive and anvi refine worked without any issues with the new multi threaded backend however anvi server started having some issues when i looked into these things i realized that the usermgmt class initiates a database connection and uses it for everything because we are relying on self contained database files and not a database server this is prone to a lot of issues going forward i changed the design a bit and created a usersdb class in the usermanagement py now usermgmt class gets a copy of this one and all database calls go through that class i started changing things just to test in usermgmt but haven t really finished i will commit my changes so far in a second but following examples the way things should look like now fetchone diff p token response self cursor execute select from users where token p user response fetchone user self users db fetchone select from users where token token fetchall diff p user response self cursor execute select name from projects where user p projects response fetchall projects self users db fetchall select name from projects where user user execute diff p token user self cursor execute update users set token where login p self conn commit self users db execute update users set token where login token user i hope this is agreeable best
0
2,698
5,541,396,341
IssuesEvent
2017-03-22 12:45:51
mathiasbynens/es-regexp-dotall-flag
https://api.github.com/repos/mathiasbynens/es-regexp-dotall-flag
closed
Advance to stage 3
process
Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] Complete spec text https://github.com/mathiasbynens/es-regexp-singleline-flag/blob/master/spec.html https://mathiasbynens.github.io/es-regexp-singleline-flag/ > - [ ] Designated reviewers have signed off on the current spec text TODO > - [ ] The ECMAScript editor has signed off on the current spec text TODO
1.0
Advance to stage 3 - Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] Complete spec text https://github.com/mathiasbynens/es-regexp-singleline-flag/blob/master/spec.html https://mathiasbynens.github.io/es-regexp-singleline-flag/ > - [ ] Designated reviewers have signed off on the current spec text TODO > - [ ] The ECMAScript editor has signed off on the current spec text TODO
process
advance to stage criteria taken from minus those from previous stages complete spec text designated reviewers have signed off on the current spec text todo the ecmascript editor has signed off on the current spec text todo
1
9,848
25,384,553,893
IssuesEvent
2022-11-21 20:38:07
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[REMOTO] Back-end developer - DB1 Group
CLT PJ Sênior PHP TDD Remoto MySQL DDD SOLID Hexagonal architecture Clean architecture
## Nossa empresa A DB1 Global Software é uma empresa do DB1 Group e é uma Software House especializada no desenvolvimento de software para grandes players do mercado. Na DGS, como chamamos aqui, trabalhamos com a transformação digital alinhada a uma entrega de valor aos nossos clientes. Temos o propósito de enfrentar desafios e resolver problemas, buscando nos aperfeiçoar cada vez mais. Aqui, entendemos que o profissional e pessoal andam juntos, por isso, cultura importa tanto quanto conhecimento técnico. Como fazemos isso? Formando squads exclusivas para nossos projetos, com equipes multidisciplinares, que imergem no segmento do nosso cliente e focam em realizar entregas pautadas em resultados e com padrão global de qualidade. Além disso, somos sempre movidos por uma cultura que desenvolve e valoriza pessoas! ## Descrição da vaga Buscamos uma pessoa apaixonada por tecnologia, com experiência em desenvolvimento backend, que se interesse em desenvolver com PHP e busque entregar soluções de qualidade aplicando as melhores práticas de programação, tais como SOLID, design patterns, clean code. Fará parte de um time de pessoas que contribui para a evolução de uma plataforma de pagamentos. ## Local Remoto ## Requisitos **Para fazer parte dessa missão é importante:** - Atuar em desenvolvimento backend PHP, liderando o time de desenvolvedores provendo solução técnica, acessando as àreas de apoio se necessário (arquitetura de solução, devops, segurança, etc), garantindo que as entregas sejam feitas com qualidade, eficiência e gerando valor para a área de negócio; - Apoiar na evolução das habilidades técnicas do time, também participará do processo de evolução de conhecimento dos desenvolvedores que lidera, garantindo evolução e aprendizado. - Fazer code reviews e ajudar o time em decisões de arquitetura. **É imprescindível que você conheça:** - PHP 7+ - Composer - MySql - Orientação a Objetos - RestFull API - PHPUnit (unitário/integração) - TDD e DDD - Conceitos de boas práticas de desenvolvimento (SOLID, GRASP, Clean Architecture, Hexagonal Architecture) - Inglês técnico para comunicação escrita (Lingua padrão utilizada pelo cliente) - Inglês avançado para conversação (participará de reuniões em inglês com cliente) - Ser uma pessoa flexível e por dentro das novidades tecnológicas; - Grande capacidade de aprendizado para novas práticas, tecnologias, linguagens de programação e culturas de engenharia. ## Benefícios - Cartão de benefícios flexíveis DUCZ (Alimentação, Mobilidade, Qualidade de Vida, entre outros); - Plano de Saúde Unimed - Com mensalidade 100% paga pela DB1; - Auxílio home office; - Horários flexíveis!; - Gympass (opcional); - Plano odontológico (opcional); - Seguro de Vida (opcional); - Evolução de carreira: Plano de Desenvolvimento Individual (PDI), feedbacks constantes, Programa de Mentoria, Parceria com Cambly e Fluency para estudo de idiomas, eventos internos, subsídio para treinamentos, universidade corporativa e muito mais; - Programas de recompensa: Next1 para indicação de novos colaboradores e bônus anual atrelado a metas; - Comitês para você participar e contribuir com a comunidade e com as melhorias dos processos do DB1 Group. ## Contratação Preferencialmente, CLT ## Como se candidatar [Candidate-se aqui](https://db1group.pinpointhq.com/pt-BR/jobs/65741?utm_medium=organic_search&utm_source=GitHub) ## Labels ### Nível - Sênior ### Regime - CLT - PJ ### Alocação - Remoto
2.0
[REMOTO] Back-end developer - DB1 Group - ## Nossa empresa A DB1 Global Software é uma empresa do DB1 Group e é uma Software House especializada no desenvolvimento de software para grandes players do mercado. Na DGS, como chamamos aqui, trabalhamos com a transformação digital alinhada a uma entrega de valor aos nossos clientes. Temos o propósito de enfrentar desafios e resolver problemas, buscando nos aperfeiçoar cada vez mais. Aqui, entendemos que o profissional e pessoal andam juntos, por isso, cultura importa tanto quanto conhecimento técnico. Como fazemos isso? Formando squads exclusivas para nossos projetos, com equipes multidisciplinares, que imergem no segmento do nosso cliente e focam em realizar entregas pautadas em resultados e com padrão global de qualidade. Além disso, somos sempre movidos por uma cultura que desenvolve e valoriza pessoas! ## Descrição da vaga Buscamos uma pessoa apaixonada por tecnologia, com experiência em desenvolvimento backend, que se interesse em desenvolver com PHP e busque entregar soluções de qualidade aplicando as melhores práticas de programação, tais como SOLID, design patterns, clean code. Fará parte de um time de pessoas que contribui para a evolução de uma plataforma de pagamentos. ## Local Remoto ## Requisitos **Para fazer parte dessa missão é importante:** - Atuar em desenvolvimento backend PHP, liderando o time de desenvolvedores provendo solução técnica, acessando as àreas de apoio se necessário (arquitetura de solução, devops, segurança, etc), garantindo que as entregas sejam feitas com qualidade, eficiência e gerando valor para a área de negócio; - Apoiar na evolução das habilidades técnicas do time, também participará do processo de evolução de conhecimento dos desenvolvedores que lidera, garantindo evolução e aprendizado. - Fazer code reviews e ajudar o time em decisões de arquitetura. **É imprescindível que você conheça:** - PHP 7+ - Composer - MySql - Orientação a Objetos - RestFull API - PHPUnit (unitário/integração) - TDD e DDD - Conceitos de boas práticas de desenvolvimento (SOLID, GRASP, Clean Architecture, Hexagonal Architecture) - Inglês técnico para comunicação escrita (Lingua padrão utilizada pelo cliente) - Inglês avançado para conversação (participará de reuniões em inglês com cliente) - Ser uma pessoa flexível e por dentro das novidades tecnológicas; - Grande capacidade de aprendizado para novas práticas, tecnologias, linguagens de programação e culturas de engenharia. ## Benefícios - Cartão de benefícios flexíveis DUCZ (Alimentação, Mobilidade, Qualidade de Vida, entre outros); - Plano de Saúde Unimed - Com mensalidade 100% paga pela DB1; - Auxílio home office; - Horários flexíveis!; - Gympass (opcional); - Plano odontológico (opcional); - Seguro de Vida (opcional); - Evolução de carreira: Plano de Desenvolvimento Individual (PDI), feedbacks constantes, Programa de Mentoria, Parceria com Cambly e Fluency para estudo de idiomas, eventos internos, subsídio para treinamentos, universidade corporativa e muito mais; - Programas de recompensa: Next1 para indicação de novos colaboradores e bônus anual atrelado a metas; - Comitês para você participar e contribuir com a comunidade e com as melhorias dos processos do DB1 Group. ## Contratação Preferencialmente, CLT ## Como se candidatar [Candidate-se aqui](https://db1group.pinpointhq.com/pt-BR/jobs/65741?utm_medium=organic_search&utm_source=GitHub) ## Labels ### Nível - Sênior ### Regime - CLT - PJ ### Alocação - Remoto
non_process
back end developer group nossa empresa a global software é uma empresa do group e é uma software house especializada no desenvolvimento de software para grandes players do mercado na dgs como chamamos aqui trabalhamos com a transformação digital alinhada a uma entrega de valor aos nossos clientes temos o propósito de enfrentar desafios e resolver problemas buscando nos aperfeiçoar cada vez mais aqui entendemos que o profissional e pessoal andam juntos por isso cultura importa tanto quanto conhecimento técnico como fazemos isso formando squads exclusivas para nossos projetos com equipes multidisciplinares que imergem no segmento do nosso cliente e focam em realizar entregas pautadas em resultados e com padrão global de qualidade além disso somos sempre movidos por uma cultura que desenvolve e valoriza pessoas descrição da vaga buscamos uma pessoa apaixonada por tecnologia com experiência em desenvolvimento backend que se interesse em desenvolver com php e busque entregar soluções de qualidade aplicando as melhores práticas de programação tais como solid design patterns clean code fará parte de um time de pessoas que contribui para a evolução de uma plataforma de pagamentos local remoto requisitos para fazer parte dessa missão é importante atuar em desenvolvimento backend php liderando o time de desenvolvedores provendo solução técnica acessando as àreas de apoio se necessário arquitetura de solução devops segurança etc garantindo que as entregas sejam feitas com qualidade eficiência e gerando valor para a área de negócio apoiar na evolução das habilidades técnicas do time também participará do processo de evolução de conhecimento dos desenvolvedores que lidera garantindo evolução e aprendizado fazer code reviews e ajudar o time em decisões de arquitetura é imprescindível que você conheça php composer mysql orientação a objetos restfull api phpunit unitário integração tdd e ddd conceitos de boas práticas de desenvolvimento solid grasp clean architecture hexagonal architecture inglês técnico para comunicação escrita lingua padrão utilizada pelo cliente inglês avançado para conversação participará de reuniões em inglês com cliente ser uma pessoa flexível e por dentro das novidades tecnológicas grande capacidade de aprendizado para novas práticas tecnologias linguagens de programação e culturas de engenharia benefícios cartão de benefícios flexíveis ducz alimentação mobilidade qualidade de vida entre outros plano de saúde unimed com mensalidade paga pela auxílio home office horários flexíveis gympass opcional plano odontológico opcional seguro de vida opcional evolução de carreira plano de desenvolvimento individual pdi feedbacks constantes programa de mentoria parceria com cambly e fluency para estudo de idiomas eventos internos subsídio para treinamentos universidade corporativa e muito mais programas de recompensa para indicação de novos colaboradores e bônus anual atrelado a metas comitês para você participar e contribuir com a comunidade e com as melhorias dos processos do group contratação preferencialmente clt como se candidatar labels nível sênior regime clt pj alocação remoto
0
222,463
24,708,940,783
IssuesEvent
2022-10-19 21:51:46
lukebrogan-mend/bag-of-holding
https://api.github.com/repos/lukebrogan-mend/bag-of-holding
closed
CVE-2020-8203 (High) detected in lodash-1.0.2.tgz, lodash-2.4.2.tgz - autoclosed
security vulnerability
## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-1.0.2.tgz</b>, <b>lodash-2.4.2.tgz</b></p></summary> <p> <details><summary><b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-sass-1.3.3.tgz (Root Library) - node-sass-2.1.1.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sass-graph/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-sass-1.3.3.tgz (Root Library) - node-sass-2.1.1.tgz - sass-graph-1.3.0.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/bag-of-holding/commit/fd67f5ea4245a25511650d4bf41a6c699a799144">fd67f5ea4245a25511650d4bf41a6c699a799144</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (gulp-sass): 2.0.0</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (gulp-sass): 2.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2020-8203 (High) detected in lodash-1.0.2.tgz, lodash-2.4.2.tgz - autoclosed - ## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-1.0.2.tgz</b>, <b>lodash-2.4.2.tgz</b></p></summary> <p> <details><summary><b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-sass-1.3.3.tgz (Root Library) - node-sass-2.1.1.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sass-graph/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-sass-1.3.3.tgz (Root Library) - node-sass-2.1.1.tgz - sass-graph-1.3.0.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/bag-of-holding/commit/fd67f5ea4245a25511650d4bf41a6c699a799144">fd67f5ea4245a25511650d4bf41a6c699a799144</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (gulp-sass): 2.0.0</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (gulp-sass): 2.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
cve high detected in lodash tgz lodash tgz autoclosed cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy gulp sass tgz root library node sass tgz gaze tgz globule tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file package json path to vulnerable library node modules sass graph node modules lodash package json dependency hierarchy gulp sass tgz root library node sass tgz sass graph tgz x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution gulp sass fix resolution lodash direct dependency fix resolution gulp sass rescue worker helmet automatic remediation is available for this issue
0
14,695
17,858,600,265
IssuesEvent
2021-09-05 14:22:27
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
how get absolute path in .bzl?
P4 type: support / not a bug (process) team-Rules-CPP
### Description of the problem / feature request: > problem ### Feature requests: what underlying problem are you trying to solve with this feature? > show "src/main/tools/process-wrapper-legacy.cc:80: execvp(toolchain/darwin_arm64e/libtool.sh, ...): No such file or directory " when i use cxx_builtin_include_directories. ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. > Replace this line with your answer. ### What operating system are you running Bazel on? > MacOs ### What's the output of `bazel info release`? > release 4.0.0-homebrew ### Have you found anything relevant by searching the web? > https://stackoverflow.com/questions/47516647/is-it-possible-to-get-pwd-value-from-bazel-bzl-rule ### Any other information, logs, or outputs that you want to share? <img width="887" alt="image" src="https://user-images.githubusercontent.com/15551108/111897861-5950ef00-8a5d-11eb-8ffc-d3b0cb025fbf.png"> <img width="1135" alt="image" src="https://user-images.githubusercontent.com/15551108/111897882-7b4a7180-8a5d-11eb-8b6f-b015fdb568fc.png">
1.0
how get absolute path in .bzl? - ### Description of the problem / feature request: > problem ### Feature requests: what underlying problem are you trying to solve with this feature? > show "src/main/tools/process-wrapper-legacy.cc:80: execvp(toolchain/darwin_arm64e/libtool.sh, ...): No such file or directory " when i use cxx_builtin_include_directories. ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. > Replace this line with your answer. ### What operating system are you running Bazel on? > MacOs ### What's the output of `bazel info release`? > release 4.0.0-homebrew ### Have you found anything relevant by searching the web? > https://stackoverflow.com/questions/47516647/is-it-possible-to-get-pwd-value-from-bazel-bzl-rule ### Any other information, logs, or outputs that you want to share? <img width="887" alt="image" src="https://user-images.githubusercontent.com/15551108/111897861-5950ef00-8a5d-11eb-8ffc-d3b0cb025fbf.png"> <img width="1135" alt="image" src="https://user-images.githubusercontent.com/15551108/111897882-7b4a7180-8a5d-11eb-8b6f-b015fdb568fc.png">
process
how get absolute path in bzl description of the problem feature request problem feature requests what underlying problem are you trying to solve with this feature show src main tools process wrapper legacy cc execvp toolchain darwin libtool sh no such file or directory when i use cxx builtin include directories bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible replace this line with your answer what operating system are you running bazel on macos what s the output of bazel info release release homebrew have you found anything relevant by searching the web any other information logs or outputs that you want to share img width alt image src img width alt image src
1
747,688
26,095,637,804
IssuesEvent
2022-12-26 19:06:08
bounswe/bounswe2022group2
https://api.github.com/repos/bounswe/bounswe2022group2
closed
Frontend: Create Event Component
priority-medium status-inprogress feature front-end
### Issue Description Events are one of the key features of our application. A user can create or participate in events for a particular learning space when they are joined to that learning space. Thus, there will be a component to create an event for the learning space in general learning space page. In our sixth meeting, we have determined the position and general look of the create event form component. I am responsible for the implementation and the further details. The create event form will feature form fields and submit button. The form fields are listed as follows: * Event title (text input) * Event description (text input) * Event date and time (date and time input) * Event duration (date and time input) * Participant limit (text input) * An interactive map input (for geolocation) ### Step Details Steps that will be performed: - [x] Search for date/time input components for React - [x] Search for map input components for React - [x] Create a general design for the form component - [x] Add title input field - [x] Add description input field - [x] Add date and time input field - [x] Add duration input field - [x] Add participant limit input field - [x] Add map input field - [x] Add form validations ### Final Actions I will create a separate branch from our frontend development branch to continue with the implementation. After everything is done and every change is committed, I will be creating a pull request accordingly and link that pull request to this issue. Upon approval and merging, the changes will be live on our development branch and this issue will be closed. ### Deadline of the Issue 22.12.2022 - Thursday - 23.59 ### Reviewer Muhammed Enes Sürmeli ### Deadline for the Review 23.12.2022 - Friday - 23.59
1.0
Frontend: Create Event Component - ### Issue Description Events are one of the key features of our application. A user can create or participate in events for a particular learning space when they are joined to that learning space. Thus, there will be a component to create an event for the learning space in general learning space page. In our sixth meeting, we have determined the position and general look of the create event form component. I am responsible for the implementation and the further details. The create event form will feature form fields and submit button. The form fields are listed as follows: * Event title (text input) * Event description (text input) * Event date and time (date and time input) * Event duration (date and time input) * Participant limit (text input) * An interactive map input (for geolocation) ### Step Details Steps that will be performed: - [x] Search for date/time input components for React - [x] Search for map input components for React - [x] Create a general design for the form component - [x] Add title input field - [x] Add description input field - [x] Add date and time input field - [x] Add duration input field - [x] Add participant limit input field - [x] Add map input field - [x] Add form validations ### Final Actions I will create a separate branch from our frontend development branch to continue with the implementation. After everything is done and every change is committed, I will be creating a pull request accordingly and link that pull request to this issue. Upon approval and merging, the changes will be live on our development branch and this issue will be closed. ### Deadline of the Issue 22.12.2022 - Thursday - 23.59 ### Reviewer Muhammed Enes Sürmeli ### Deadline for the Review 23.12.2022 - Friday - 23.59
non_process
frontend create event component issue description events are one of the key features of our application a user can create or participate in events for a particular learning space when they are joined to that learning space thus there will be a component to create an event for the learning space in general learning space page in our sixth meeting we have determined the position and general look of the create event form component i am responsible for the implementation and the further details the create event form will feature form fields and submit button the form fields are listed as follows event title text input event description text input event date and time date and time input event duration date and time input participant limit text input an interactive map input for geolocation step details steps that will be performed search for date time input components for react search for map input components for react create a general design for the form component add title input field add description input field add date and time input field add duration input field add participant limit input field add map input field add form validations final actions i will create a separate branch from our frontend development branch to continue with the implementation after everything is done and every change is committed i will be creating a pull request accordingly and link that pull request to this issue upon approval and merging the changes will be live on our development branch and this issue will be closed deadline of the issue thursday reviewer muhammed enes sürmeli deadline for the review friday
0
507,321
14,679,965,562
IssuesEvent
2020-12-31 08:38:10
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/docs/reference/glossary/kube-apiserver.md
lang/zh priority/P0 sync/update version/master welcome
Source File: [/docs/reference/glossary/kube-apiserver.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/glossary/kube-apiserver.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/glossary/kube-apiserver.md content/zh/docs/reference/glossary/kube-apiserver.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/glossary/kube-apiserver.md ```
1.0
/docs/reference/glossary/kube-apiserver.md - Source File: [/docs/reference/glossary/kube-apiserver.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/glossary/kube-apiserver.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/glossary/kube-apiserver.md content/zh/docs/reference/glossary/kube-apiserver.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/glossary/kube-apiserver.md ```
non_process
docs reference glossary kube apiserver md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs reference glossary kube apiserver md content zh docs reference glossary kube apiserver md 跨分支持查看原始文档更新差异 git diff release master content en docs reference glossary kube apiserver md
0
9,985
13,033,759,303
IssuesEvent
2020-07-28 07:33:35
crystaluniverse/publishingtools
https://api.github.com/repos/crystaluniverse/publishingtools
closed
error when build tool with crystal 0.34.0
process_wontfix
``` codescalerssMBP:publishingtools codescalers$ bash build.sh ``` ![image](https://user-images.githubusercontent.com/21987477/85050309-2dcf6a00-b196-11ea-95d0-7abf3e04587f.png) ``` codescalerssMBP:publishingtools codescalers$ crystal --version Crystal 0.34.0 (2020-04-06) LLVM: 6.0.1 Default target: x86_64-apple-macosx ```
1.0
error when build tool with crystal 0.34.0 - ``` codescalerssMBP:publishingtools codescalers$ bash build.sh ``` ![image](https://user-images.githubusercontent.com/21987477/85050309-2dcf6a00-b196-11ea-95d0-7abf3e04587f.png) ``` codescalerssMBP:publishingtools codescalers$ crystal --version Crystal 0.34.0 (2020-04-06) LLVM: 6.0.1 Default target: x86_64-apple-macosx ```
process
error when build tool with crystal codescalerssmbp publishingtools codescalers bash build sh codescalerssmbp publishingtools codescalers crystal version crystal llvm default target apple macosx
1
217,752
7,327,925,368
IssuesEvent
2018-03-04 15:43:30
openSUSE/docbookrx
https://api.github.com/repos/openSUSE/docbookrx
closed
Add Appendix as a supported document type
High Priority enhancement
Appendix is missing from the supported document types: `/lib/docbookrx/docbook_visitor.rb` This is required as some of SUSE's larger books have appendix documents and do not just use the tag internal to a document. The document itself is an appendix and is linked from a book entry. ```Ruby DOCUMENT_NAMES = ['article', 'book'] ```
1.0
Add Appendix as a supported document type - Appendix is missing from the supported document types: `/lib/docbookrx/docbook_visitor.rb` This is required as some of SUSE's larger books have appendix documents and do not just use the tag internal to a document. The document itself is an appendix and is linked from a book entry. ```Ruby DOCUMENT_NAMES = ['article', 'book'] ```
non_process
add appendix as a supported document type appendix is missing from the supported document types lib docbookrx docbook visitor rb this is required as some of suse s larger books have appendix documents and do not just use the tag internal to a document the document itself is an appendix and is linked from a book entry ruby document names
0
207,957
23,529,511,376
IssuesEvent
2022-08-19 14:04:19
MatBenfield/news
https://api.github.com/repos/MatBenfield/news
closed
[SecurityWeek] Evasive 'DarkTortilla' Crypter Delivers RATs, Targeted Malware
SecurityWeek Stale
**Secureworks security researchers have analyzed ‘DarkTortilla’, a .NET-based crypter used to deliver both popular malware and targeted payloads.** [read more](https://www.securityweek.com/evasive-darktortilla-crypter-delivers-rats-targeted-malware) <https://www.securityweek.com/evasive-darktortilla-crypter-delivers-rats-targeted-malware>
True
[SecurityWeek] Evasive 'DarkTortilla' Crypter Delivers RATs, Targeted Malware - **Secureworks security researchers have analyzed ‘DarkTortilla’, a .NET-based crypter used to deliver both popular malware and targeted payloads.** [read more](https://www.securityweek.com/evasive-darktortilla-crypter-delivers-rats-targeted-malware) <https://www.securityweek.com/evasive-darktortilla-crypter-delivers-rats-targeted-malware>
non_process
evasive darktortilla crypter delivers rats targeted malware secureworks security researchers have analyzed ‘darktortilla’ a net based crypter used to deliver both popular malware and targeted payloads
0
15,880
20,070,548,980
IssuesEvent
2022-02-04 05:56:40
swig/swig
https://api.github.com/repos/swig/swig
opened
-DFOO works different to a real compiler preprocessor
preprocessor
In a C/C++ compiler, `-DFOO` on the command line sets `FOO` to `1`. In SWIG it sets `FOO` to an empty value. (In both SWIG and compilers, `#define FOO` in a file sets `FOO` to an empty value.) Reproducer (same result with 4.0.2 and git master): ``` $ cat emptydefine.i #define EMPTY2 [EMPTY1] [EMPTY2] $ gcc -E -xc -DEMPTY1 emptydefine.i # 0 "emptydefine.i" # 0 "<built-in>" # 0 "<command-line>" # 1 "/usr/include/stdc-predef.h" 1 3 4 # 0 "<command-line>" 2 # 1 "emptydefine.i" [1] [] $ clang-13 -E -xc -DEMPTY1 emptydefine.i # 1 "emptydefine.i" # 1 "<built-in>" 1 # 1 "<built-in>" 3 # 343 "<built-in>" 3 # 1 "<command line>" 1 # 1 "<built-in>" 2 # 1 "emptydefine.i" 2 [1] [] $ swig -E -python -DEMPTY1 emptydefine.i |tail -n3 [] [] %endoffile ``` The only SWIG documentation for `-D` seems to be this which says nothing about the value (not even that `-DFOO=bar` is supported): > -D<symbol> - Define a symbol <symbol> (for conditional compilation) Overall it seems better to fix this to me. It goes have a slight potential to break existing use of SWIG, but it's easy to fix - `-DFOO=` if you really want an empty value works now and would continue to work), but it would mean expectations based on compilers would be met going forwards (this quirk just tripped me up while trying to test something). 4.1.0 seems a reasonable release to address this in.
1.0
-DFOO works different to a real compiler preprocessor - In a C/C++ compiler, `-DFOO` on the command line sets `FOO` to `1`. In SWIG it sets `FOO` to an empty value. (In both SWIG and compilers, `#define FOO` in a file sets `FOO` to an empty value.) Reproducer (same result with 4.0.2 and git master): ``` $ cat emptydefine.i #define EMPTY2 [EMPTY1] [EMPTY2] $ gcc -E -xc -DEMPTY1 emptydefine.i # 0 "emptydefine.i" # 0 "<built-in>" # 0 "<command-line>" # 1 "/usr/include/stdc-predef.h" 1 3 4 # 0 "<command-line>" 2 # 1 "emptydefine.i" [1] [] $ clang-13 -E -xc -DEMPTY1 emptydefine.i # 1 "emptydefine.i" # 1 "<built-in>" 1 # 1 "<built-in>" 3 # 343 "<built-in>" 3 # 1 "<command line>" 1 # 1 "<built-in>" 2 # 1 "emptydefine.i" 2 [1] [] $ swig -E -python -DEMPTY1 emptydefine.i |tail -n3 [] [] %endoffile ``` The only SWIG documentation for `-D` seems to be this which says nothing about the value (not even that `-DFOO=bar` is supported): > -D<symbol> - Define a symbol <symbol> (for conditional compilation) Overall it seems better to fix this to me. It goes have a slight potential to break existing use of SWIG, but it's easy to fix - `-DFOO=` if you really want an empty value works now and would continue to work), but it would mean expectations based on compilers would be met going forwards (this quirk just tripped me up while trying to test something). 4.1.0 seems a reasonable release to address this in.
process
dfoo works different to a real compiler preprocessor in a c c compiler dfoo on the command line sets foo to in swig it sets foo to an empty value in both swig and compilers define foo in a file sets foo to an empty value reproducer same result with and git master cat emptydefine i define gcc e xc emptydefine i emptydefine i usr include stdc predef h emptydefine i clang e xc emptydefine i emptydefine i emptydefine i swig e python emptydefine i tail endoffile the only swig documentation for d seems to be this which says nothing about the value not even that dfoo bar is supported d define a symbol for conditional compilation overall it seems better to fix this to me it goes have a slight potential to break existing use of swig but it s easy to fix dfoo if you really want an empty value works now and would continue to work but it would mean expectations based on compilers would be met going forwards this quirk just tripped me up while trying to test something seems a reasonable release to address this in
1
3,061
6,047,932,293
IssuesEvent
2017-06-12 15:25:01
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
reopened
claim asd failed but gui alert is claimed
process_moved
Unable to claim an asd but gui shows claimed. After refreshing the page the asds is still available to claim. ![claimed](https://user-images.githubusercontent.com/2900612/27035520-e71d0c30-4f81-11e7-8d3b-3f6891263162.png) Error in the logging: ``` Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 13200 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182759 - WARNING - 2017-06-12 07:20:53 13200 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182758 - DEBUG - stdout: {"success":false,"error":{"message":"Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure \"Albamgr: too many retries for operation MarkOsdClaimed\"))","exception_type":"albamgr_exn","exception_code":1}} Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 25100 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182761 - WARNING - 2017-06-12 07:20:53 25000 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182760 - ERROR - Error: Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure "Albamgr: too many retries for operation MarkOsdClaimed")) Jun 12 07:20:53 NY1SRV0019 celery[13127]: Traceback (most recent call last): Jun 12 07:20:53 NY1SRV0019 celery[13127]: File "/opt/OpenvStorage/ovs/extensions/plugins/albacli.py", line 139, in run Jun 12 07:20:53 NY1SRV0019 celery[13127]: raise AlbaError(output['error']['message'], output['error']['exception_code']) Jun 12 07:20:53 NY1SRV0019 celery[13127]: AlbaError: Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure "Albamgr: too many retries for operation MarkOsdClaimed")) Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 25200 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182763 - WARNING - 2017-06-12 07:20:53 25200 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182762 - DEBUG - Command: /usr/bin/alba claim-osd --to-json --config=arakoon://config/ovs/arakoon/ny1-ssdbackend01-abm/config?ini=%2Fopt%2FOpenvStorage%2Fconfig%2Farakoon_cacc.ini --long-id=H5H5Nh0qINuZuHR11gCJDhHFFx7nxAbN ``` unable to delete the disk: ![unable_to_delete](https://user-images.githubusercontent.com/2900612/27036375-f827ebf0-4f84-11e7-81f6-51f08ca355a5.png) asd is not registered in the ABM: ``` root@NY1SRV0001:~# alba list-osds --config arakoon://config/ovs/arakoon/ny1-ssdbackend01-abm/config?ini=%2Fmnt%2Fssd1%2Farakoon%2Fexternal_arakoon_cacc.ini --to-json | grep H5H 2017-06-12 09:47:37 237048 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 0 - info - Albamgr_client.make_client :ny1-ssdbackend01-abm 2017-06-12 09:47:37 238037 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 1 - info - connect_with : 172.17.16.5 26406 None Net_fd.TCP (fd:3) 2017-06-12 09:47:37 265195 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 2 - info - closing (fd:3) root@NY1SRV0001:~# ```
1.0
claim asd failed but gui alert is claimed - Unable to claim an asd but gui shows claimed. After refreshing the page the asds is still available to claim. ![claimed](https://user-images.githubusercontent.com/2900612/27035520-e71d0c30-4f81-11e7-8d3b-3f6891263162.png) Error in the logging: ``` Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 13200 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182759 - WARNING - 2017-06-12 07:20:53 13200 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182758 - DEBUG - stdout: {"success":false,"error":{"message":"Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure \"Albamgr: too many retries for operation MarkOsdClaimed\"))","exception_type":"albamgr_exn","exception_code":1}} Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 25100 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182761 - WARNING - 2017-06-12 07:20:53 25000 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182760 - ERROR - Error: Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure "Albamgr: too many retries for operation MarkOsdClaimed")) Jun 12 07:20:53 NY1SRV0019 celery[13127]: Traceback (most recent call last): Jun 12 07:20:53 NY1SRV0019 celery[13127]: File "/opt/OpenvStorage/ovs/extensions/plugins/albacli.py", line 139, in run Jun 12 07:20:53 NY1SRV0019 celery[13127]: raise AlbaError(output['error']['message'], output['error']['exception_code']) Jun 12 07:20:53 NY1SRV0019 celery[13127]: AlbaError: Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,(Failure "Albamgr: too many retries for operation MarkOsdClaimed")) Jun 12 07:20:53 NY1SRV0019 celery[13127]: 2017-06-12 07:20:53 25200 -0400 - NY1SRV0019 - 14977/139952705304320 - celery/celery.redirected - 182763 - WARNING - 2017-06-12 07:20:53 25200 -0400 - NY1SRV0019 - 14977/139952705304320 - extensions/alba-cli - 182762 - DEBUG - Command: /usr/bin/alba claim-osd --to-json --config=arakoon://config/ovs/arakoon/ny1-ssdbackend01-abm/config?ini=%2Fopt%2FOpenvStorage%2Fconfig%2Farakoon_cacc.ini --long-id=H5H5Nh0qINuZuHR11gCJDhHFFx7nxAbN ``` unable to delete the disk: ![unable_to_delete](https://user-images.githubusercontent.com/2900612/27036375-f827ebf0-4f84-11e7-81f6-51f08ca355a5.png) asd is not registered in the ABM: ``` root@NY1SRV0001:~# alba list-osds --config arakoon://config/ovs/arakoon/ny1-ssdbackend01-abm/config?ini=%2Fmnt%2Fssd1%2Farakoon%2Fexternal_arakoon_cacc.ini --to-json | grep H5H 2017-06-12 09:47:37 237048 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 0 - info - Albamgr_client.make_client :ny1-ssdbackend01-abm 2017-06-12 09:47:37 238037 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 1 - info - connect_with : 172.17.16.5 26406 None Net_fd.TCP (fd:3) 2017-06-12 09:47:37 265195 -0400 - NY1SRV0001 - 2441/0000 - alba/cli - 2 - info - closing (fd:3) root@NY1SRV0001:~# ```
process
claim asd failed but gui alert is claimed unable to claim an asd but gui shows claimed after refreshing the page the asds is still available to claim error in the logging jun celery celery celery redirected warning extensions alba cli debug stdout success false error message albamgr exception albamgr protocol protocol error unknown failure albamgr too many retries for operation markosdclaimed exception type albamgr exn exception code jun celery celery celery redirected warning extensions alba cli error error albamgr exception albamgr protocol protocol error unknown failure albamgr too many retries for operation markosdclaimed jun celery traceback most recent call last jun celery file opt openvstorage ovs extensions plugins albacli py line in run jun celery raise albaerror output output jun celery albaerror albamgr exception albamgr protocol protocol error unknown failure albamgr too many retries for operation markosdclaimed jun celery celery celery redirected warning extensions alba cli debug command usr bin alba claim osd to json config arakoon config ovs arakoon abm config ini cacc ini long id unable to delete the disk asd is not registered in the abm root alba list osds config arakoon config ovs arakoon abm config ini arakoon cacc ini to json grep alba cli info albamgr client make client abm alba cli info connect with none net fd tcp fd alba cli info closing fd root
1
58
2,518,635,051
IssuesEvent
2015-01-17 00:06:33
MozillaFoundation/plan
https://api.github.com/repos/MozillaFoundation/plan
closed
ENGINEERING: Create reference app using stack from engineering handbook; communicate to team
p1 process
We need a reference application we can all refer to when suggesting changes to our stack. Phase: Build / Ship Owner: @simonwex Decision: @simonwex Lead design: n/a Lead dev: @Pomax Quality: @jbuck Dev: @alicoding
1.0
ENGINEERING: Create reference app using stack from engineering handbook; communicate to team - We need a reference application we can all refer to when suggesting changes to our stack. Phase: Build / Ship Owner: @simonwex Decision: @simonwex Lead design: n/a Lead dev: @Pomax Quality: @jbuck Dev: @alicoding
process
engineering create reference app using stack from engineering handbook communicate to team we need a reference application we can all refer to when suggesting changes to our stack phase build ship owner simonwex decision simonwex lead design n a lead dev pomax quality jbuck dev alicoding
1
3,724
6,732,910,472
IssuesEvent
2017-10-18 13:16:11
lockedata/rcms
https://api.github.com/repos/lockedata/rcms
opened
Manage registration
attendee odoo processes
## Detailed task - Edit your information - Get a refund on your ticker ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [odoo](//http://188.166.159.192:8069) - System documentation: [odoo docs](https://www.odoo.com/page/docs) - Role: Attendee - Area: Processes
1.0
Manage registration - ## Detailed task - Edit your information - Get a refund on your ticker ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [odoo](//http://188.166.159.192:8069) - System documentation: [odoo docs](https://www.odoo.com/page/docs) - Role: Attendee - Area: Processes
process
manage registration detailed task edit your information get a refund on your ticker assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role attendee area processes
1
18,053
24,065,593,101
IssuesEvent
2022-09-17 12:43:43
tradingstrategy-ai/frontend
https://api.github.com/repos/tradingstrategy-ai/frontend
opened
Replace Cypress testing framework with a more modern alternative
priority: P2 process
As discussed in the engineering retrospective, the current testing framework is showing its age. It should be replaced with a newer framework that makes the tests easier to write and faster to execute. P.S.: When @hieuh25 creates size estimate labels in all repos, this should likely be labeled as `size: M`, as it can be probably done in a week. See the existing [size labels](https://github.com/tradingstrategy-ai/backend/labels?q=size) in the `backend` repo for reference.
1.0
Replace Cypress testing framework with a more modern alternative - As discussed in the engineering retrospective, the current testing framework is showing its age. It should be replaced with a newer framework that makes the tests easier to write and faster to execute. P.S.: When @hieuh25 creates size estimate labels in all repos, this should likely be labeled as `size: M`, as it can be probably done in a week. See the existing [size labels](https://github.com/tradingstrategy-ai/backend/labels?q=size) in the `backend` repo for reference.
process
replace cypress testing framework with a more modern alternative as discussed in the engineering retrospective the current testing framework is showing its age it should be replaced with a newer framework that makes the tests easier to write and faster to execute p s when creates size estimate labels in all repos this should likely be labeled as size m as it can be probably done in a week see the existing in the backend repo for reference
1
152,824
13,484,240,492
IssuesEvent
2020-09-11 05:57:13
rootsj/FinPic
https://api.github.com/repos/rootsj/FinPic
closed
Vue.dev1.01 update issue
documentation
Vue.dev1.01 I did merge my local Vue project into Vue.dev1.01 directly without any pull request. Update History 1. like button fix 2. report button fix 3. my page user view fix plz check it by yourself and leave your comment
1.0
Vue.dev1.01 update issue - Vue.dev1.01 I did merge my local Vue project into Vue.dev1.01 directly without any pull request. Update History 1. like button fix 2. report button fix 3. my page user view fix plz check it by yourself and leave your comment
non_process
vue update issue vue i did merge my local vue project into vue directly without any pull request update history like button fix report button fix my page user view fix plz check it by yourself and leave your comment
0
19,545
25,866,121,714
IssuesEvent
2022-12-13 21:04:00
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
opened
Processing should put operation after stem title of inputs
type - enhancement level - easy f - computations f - user-interface f - processing
i.e. "My Sample FFT" instead of "FFT of My Sample" See - #32
1.0
Processing should put operation after stem title of inputs - i.e. "My Sample FFT" instead of "FFT of My Sample" See - #32
process
processing should put operation after stem title of inputs i e my sample fft instead of fft of my sample see
1
587,612
17,620,661,514
IssuesEvent
2021-08-18 14:56:50
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
[dashboard] The behavior between the fields is different
Type:Enhancement Priority:Low Team:Frontend Status:Analysed
The `Value` and `No. Records` fields do not accept negative values. But the behavior between the fields is different. - **Value**: the minimum value is zero - **Nº Records**: the value below zero is "1" **Step to reproduce the problem**: 1. click on the down arrow in the 2 fields ![Screenshot_32](https://user-images.githubusercontent.com/37310063/100006575-758f9880-2da9-11eb-8dbb-02a10fa40c6a.png) Affected baseline: v0.5.0-rc.2
1.0
[dashboard] The behavior between the fields is different - The `Value` and `No. Records` fields do not accept negative values. But the behavior between the fields is different. - **Value**: the minimum value is zero - **Nº Records**: the value below zero is "1" **Step to reproduce the problem**: 1. click on the down arrow in the 2 fields ![Screenshot_32](https://user-images.githubusercontent.com/37310063/100006575-758f9880-2da9-11eb-8dbb-02a10fa40c6a.png) Affected baseline: v0.5.0-rc.2
non_process
the behavior between the fields is different the value and no records fields do not accept negative values but the behavior between the fields is different value the minimum value is zero nº records the value below zero is step to reproduce the problem click on the down arrow in the fields affected baseline rc
0
426,165
12,368,570,763
IssuesEvent
2020-05-18 14:01:44
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
meet.google.com - see bug description
browser-firefox engine-gecko priority-critical
<!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53065 --> **URL**: https://meet.google.com/qze-tjzw-ozs?authuser=0 **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Other **Problem type**: Something else **Description**: kiking me out for no reason **Steps to Reproduce**: i joined and then google meet kicked me out for seemingly no reason. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/c2e2c4eb-fae6-4381-bcd7-3a723e390138.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200517213618</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/55557fd3-9899-46e6-9619-512406fc6f99) Submitted in the name of `@anoymus` _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
meet.google.com - see bug description - <!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53065 --> **URL**: https://meet.google.com/qze-tjzw-ozs?authuser=0 **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Other **Problem type**: Something else **Description**: kiking me out for no reason **Steps to Reproduce**: i joined and then google meet kicked me out for seemingly no reason. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/c2e2c4eb-fae6-4381-bcd7-3a723e390138.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200517213618</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/55557fd3-9899-46e6-9619-512406fc6f99) Submitted in the name of `@anoymus` _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
meet google com see bug description url browser version firefox operating system windows tested another browser yes other problem type something else description kiking me out for no reason steps to reproduce i joined and then google meet kicked me out for seemingly no reason view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of anoymus from with ❤️
0
113,739
14,482,648,287
IssuesEvent
2020-12-10 14:14:31
phetsims/ratio-and-proportion
https://api.github.com/repos/phetsims/ratio-and-proportion
opened
Explore and verify best placement of position information for slider interactions
design:a11y
While testing #245 and #261, I found the flipped design was implemented as I requested. That said, I want to note that when I use the slider interactions (individual left or right hands), I feel like I need position as part of the object response. This is not the case with the Both Hands interaction where I receive comprehensive position information for each hand while interacting. This is an interesting design issue. I am opening this issue for future investigation. It likely won't happen for the publication for this version. At Tuesday (Dec 8th) design meeting, I also mentioned that I felt like I was missing position information sometimes. This issue can explore that "possible" design gap further as we move the current version of the sim towards publication.
1.0
Explore and verify best placement of position information for slider interactions - While testing #245 and #261, I found the flipped design was implemented as I requested. That said, I want to note that when I use the slider interactions (individual left or right hands), I feel like I need position as part of the object response. This is not the case with the Both Hands interaction where I receive comprehensive position information for each hand while interacting. This is an interesting design issue. I am opening this issue for future investigation. It likely won't happen for the publication for this version. At Tuesday (Dec 8th) design meeting, I also mentioned that I felt like I was missing position information sometimes. This issue can explore that "possible" design gap further as we move the current version of the sim towards publication.
non_process
explore and verify best placement of position information for slider interactions while testing and i found the flipped design was implemented as i requested that said i want to note that when i use the slider interactions individual left or right hands i feel like i need position as part of the object response this is not the case with the both hands interaction where i receive comprehensive position information for each hand while interacting this is an interesting design issue i am opening this issue for future investigation it likely won t happen for the publication for this version at tuesday dec design meeting i also mentioned that i felt like i was missing position information sometimes this issue can explore that possible design gap further as we move the current version of the sim towards publication
0
598,331
18,242,762,640
IssuesEvent
2021-10-01 14:41:29
BeccaLyria/discord-bot
https://api.github.com/repos/BeccaLyria/discord-bot
closed
[CHORE] - Update Message Logs
🟩 priority: low 🏁 status: ready for dev ✨ goal: improvement 🕹 aspect: interface good first issue Hacktoberfest
### What work needs to be performed? Update the message deletion and message edit logs to include the channel name, and the edit logs to include the message link. To complete this: - In [this file](https://github.com/BeccaLyria/discord-bot/blob/main/src/events/messageEvents/messageUpdate.ts) add a field to the `updateEmbed` with the name `Channel` and the value `<#${newMessage.channel.id}>`. - Add another field to that `updateEmbed` with the name `Message link` and the value `newMessage.url`. - In [this file](https://github.com/BeccaLyria/discord-bot/blob/main/src/events/messageEvents/messageDelete.ts) add a field to the `deleteEmbed` with the name `Channel` and the value `<#${channel.id}>`. ### Additional information _No response_
1.0
[CHORE] - Update Message Logs - ### What work needs to be performed? Update the message deletion and message edit logs to include the channel name, and the edit logs to include the message link. To complete this: - In [this file](https://github.com/BeccaLyria/discord-bot/blob/main/src/events/messageEvents/messageUpdate.ts) add a field to the `updateEmbed` with the name `Channel` and the value `<#${newMessage.channel.id}>`. - Add another field to that `updateEmbed` with the name `Message link` and the value `newMessage.url`. - In [this file](https://github.com/BeccaLyria/discord-bot/blob/main/src/events/messageEvents/messageDelete.ts) add a field to the `deleteEmbed` with the name `Channel` and the value `<#${channel.id}>`. ### Additional information _No response_
non_process
update message logs what work needs to be performed update the message deletion and message edit logs to include the channel name and the edit logs to include the message link to complete this in add a field to the updateembed with the name channel and the value add another field to that updateembed with the name message link and the value newmessage url in add a field to the deleteembed with the name channel and the value additional information no response
0
709,575
24,383,504,254
IssuesEvent
2022-10-04 09:47:12
strusoft/femdesign-api
https://api.github.com/repos/strusoft/femdesign-api
closed
Pipe: Method to run a script and wait for it to finish
priority:next
### Goal Something like this ```c# // namespace FemDesign // class ApplicationConnection public void RunScript(FdScript fdScript) ``` I think the name `FemDesign.ApplicationConnection` is explicit enough. **Usage** ```c# using (var pipe = new FemDesign.ApplicationConnection()) { pipe.RunScript(fdScript); } ``` ### Reference When testing (#464) we created the following code to handle running and waiting for it to finished: ```c# pipe.Send($"run {Path.GetFullPath(fdScript.FdScriptPath)}"); var guid = Guid.NewGuid(); pipe.Send($"echo {guid}"); bool isDone = false; void onOutput(string output) { if (output == guid.ToString()) { isDone = true; pipe.OnOutput -= onOutput; } }; pipe.OnOutput += onOutput; while (!isDone) { Thread.Sleep(10); } ```
1.0
Pipe: Method to run a script and wait for it to finish - ### Goal Something like this ```c# // namespace FemDesign // class ApplicationConnection public void RunScript(FdScript fdScript) ``` I think the name `FemDesign.ApplicationConnection` is explicit enough. **Usage** ```c# using (var pipe = new FemDesign.ApplicationConnection()) { pipe.RunScript(fdScript); } ``` ### Reference When testing (#464) we created the following code to handle running and waiting for it to finished: ```c# pipe.Send($"run {Path.GetFullPath(fdScript.FdScriptPath)}"); var guid = Guid.NewGuid(); pipe.Send($"echo {guid}"); bool isDone = false; void onOutput(string output) { if (output == guid.ToString()) { isDone = true; pipe.OnOutput -= onOutput; } }; pipe.OnOutput += onOutput; while (!isDone) { Thread.Sleep(10); } ```
non_process
pipe method to run a script and wait for it to finish goal something like this c namespace femdesign class applicationconnection public void runscript fdscript fdscript i think the name femdesign applicationconnection is explicit enough usage c using var pipe new femdesign applicationconnection pipe runscript fdscript reference when testing we created the following code to handle running and waiting for it to finished c pipe send run path getfullpath fdscript fdscriptpath var guid guid newguid pipe send echo guid bool isdone false void onoutput string output if output guid tostring isdone true pipe onoutput onoutput pipe onoutput onoutput while isdone thread sleep
0
1,586
6,572,360,079
IssuesEvent
2017-09-11 01:42:22
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
datadog_event module does not correctly set host for event
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME datadog_event ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### SUMMARY The `datadog_event` module does not allow correctly setting the host to be associated with events. The hostname should be passed to Datadog using the 'host' parameter in their API. The documentation for this module shows setting a tag called 'host', which does not achieve the same thing. ##### STEPS TO REPRODUCE Use the `datadog_event` module to submit an event to Datadog indicating that an Ansible run has completed, for example. As indicated in the documentation, set a host tag, e.g. `tags="host:example.com"` In Datadog, verify that the event appears in the event stream. And shows the host tag that was set. Now search for `host:example.com`. The event just submitted does not appear, because it doesn't match the search (!). This is certainly an confusing quirk in the Datadog interface, but their support team insist that the correct way to submit a host tag is using the 'host' parameter in their API, which this Ansible module does not currently support. ##### EXPECTED RESULTS I would expect the submitted Datadog event to be correctly associated with the host. The problem is a combination of two things - primarily this module's lack of support for setting the 'host' parameter in the call to Datadog's API. The second is the misleading documentation, which shows setting a 'host' tag using the 'tags' parameter. And this probably came about because the Datadog documentation for this API method does not demonstrate use of the 'host' parameter in the example code.
True
datadog_event module does not correctly set host for event - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME datadog_event ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### SUMMARY The `datadog_event` module does not allow correctly setting the host to be associated with events. The hostname should be passed to Datadog using the 'host' parameter in their API. The documentation for this module shows setting a tag called 'host', which does not achieve the same thing. ##### STEPS TO REPRODUCE Use the `datadog_event` module to submit an event to Datadog indicating that an Ansible run has completed, for example. As indicated in the documentation, set a host tag, e.g. `tags="host:example.com"` In Datadog, verify that the event appears in the event stream. And shows the host tag that was set. Now search for `host:example.com`. The event just submitted does not appear, because it doesn't match the search (!). This is certainly an confusing quirk in the Datadog interface, but their support team insist that the correct way to submit a host tag is using the 'host' parameter in their API, which this Ansible module does not currently support. ##### EXPECTED RESULTS I would expect the submitted Datadog event to be correctly associated with the host. The problem is a combination of two things - primarily this module's lack of support for setting the 'host' parameter in the call to Datadog's API. The second is the misleading documentation, which shows setting a 'host' tag using the 'tags' parameter. And this probably came about because the Datadog documentation for this API method does not demonstrate use of the 'host' parameter in the example code.
non_process
datadog event module does not correctly set host for event issue type bug report component name datadog event ansible version ansible summary the datadog event module does not allow correctly setting the host to be associated with events the hostname should be passed to datadog using the host parameter in their api the documentation for this module shows setting a tag called host which does not achieve the same thing steps to reproduce use the datadog event module to submit an event to datadog indicating that an ansible run has completed for example as indicated in the documentation set a host tag e g tags host example com in datadog verify that the event appears in the event stream and shows the host tag that was set now search for host example com the event just submitted does not appear because it doesn t match the search this is certainly an confusing quirk in the datadog interface but their support team insist that the correct way to submit a host tag is using the host parameter in their api which this ansible module does not currently support expected results i would expect the submitted datadog event to be correctly associated with the host the problem is a combination of two things primarily this module s lack of support for setting the host parameter in the call to datadog s api the second is the misleading documentation which shows setting a host tag using the tags parameter and this probably came about because the datadog documentation for this api method does not demonstrate use of the host parameter in the example code
0
53,748
13,262,231,739
IssuesEvent
2020-08-20 21:21:26
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
SN HitSpool offline analysis pages need to be reactivated (Trac #2003)
Migrated from Trac analysis defect
The automatic offline analysis of SNALERT HitSpool data on convey.icecube.wisc.edu has not been running for about a year. This was set up by David Heereman and the data and scripts are on convey but are not being processed. (Perhaps a cronjob was deleted.) This needs a few days of TLC to get up and running. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2003">https://code.icecube.wisc.edu/projects/icecube/ticket/2003</a>, reported by sybenzviand owned by afritz</em></summary> <p> ```json { "status": "closed", "changetime": "2019-04-23T09:10:16", "_ts": "1556010616958122", "description": "The automatic offline analysis of SNALERT HitSpool data on convey.icecube.wisc.edu has not been running for about a year. This was set up by David Heereman and the data and scripts are on convey but are not being processed. (Perhaps a cronjob was deleted.)\n\nThis needs a few days of TLC to get up and running.", "reporter": "sybenzvi", "cc": "jbraun, rcross, gmoment", "resolution": "fixed", "time": "2017-05-05T20:46:35", "component": "analysis", "summary": "SN HitSpool offline analysis pages need to be reactivated", "priority": "normal", "keywords": "sndaq, hitspool", "milestone": "Summer Solstice 2019", "owner": "afritz", "type": "defect" } ``` </p> </details>
1.0
SN HitSpool offline analysis pages need to be reactivated (Trac #2003) - The automatic offline analysis of SNALERT HitSpool data on convey.icecube.wisc.edu has not been running for about a year. This was set up by David Heereman and the data and scripts are on convey but are not being processed. (Perhaps a cronjob was deleted.) This needs a few days of TLC to get up and running. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2003">https://code.icecube.wisc.edu/projects/icecube/ticket/2003</a>, reported by sybenzviand owned by afritz</em></summary> <p> ```json { "status": "closed", "changetime": "2019-04-23T09:10:16", "_ts": "1556010616958122", "description": "The automatic offline analysis of SNALERT HitSpool data on convey.icecube.wisc.edu has not been running for about a year. This was set up by David Heereman and the data and scripts are on convey but are not being processed. (Perhaps a cronjob was deleted.)\n\nThis needs a few days of TLC to get up and running.", "reporter": "sybenzvi", "cc": "jbraun, rcross, gmoment", "resolution": "fixed", "time": "2017-05-05T20:46:35", "component": "analysis", "summary": "SN HitSpool offline analysis pages need to be reactivated", "priority": "normal", "keywords": "sndaq, hitspool", "milestone": "Summer Solstice 2019", "owner": "afritz", "type": "defect" } ``` </p> </details>
non_process
sn hitspool offline analysis pages need to be reactivated trac the automatic offline analysis of snalert hitspool data on convey icecube wisc edu has not been running for about a year this was set up by david heereman and the data and scripts are on convey but are not being processed perhaps a cronjob was deleted this needs a few days of tlc to get up and running migrated from json status closed changetime ts description the automatic offline analysis of snalert hitspool data on convey icecube wisc edu has not been running for about a year this was set up by david heereman and the data and scripts are on convey but are not being processed perhaps a cronjob was deleted n nthis needs a few days of tlc to get up and running reporter sybenzvi cc jbraun rcross gmoment resolution fixed time component analysis summary sn hitspool offline analysis pages need to be reactivated priority normal keywords sndaq hitspool milestone summer solstice owner afritz type defect
0
14,079
16,959,806,039
IssuesEvent
2021-06-29 00:59:45
carbon-design-system/ibm-cloud-cognitive
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
closed
Updates to the ReadMe
help wanted status: wontfix type: process improvement
Could the ReadMe use an update? I see it still references “Experimental” components. For a developer coming in for the first time and consuming the components in their product, do we have instructions for installing and examples of importing and using a component? And any other info they may need to get started consuming. I'm thinking we can follow [carbon-react's ReadMe](https://github.com/carbon-design-system/carbon/tree/master/packages/react#carbon-components-react)
1.0
Updates to the ReadMe - Could the ReadMe use an update? I see it still references “Experimental” components. For a developer coming in for the first time and consuming the components in their product, do we have instructions for installing and examples of importing and using a component? And any other info they may need to get started consuming. I'm thinking we can follow [carbon-react's ReadMe](https://github.com/carbon-design-system/carbon/tree/master/packages/react#carbon-components-react)
process
updates to the readme could the readme use an update i see it still references “experimental” components for a developer coming in for the first time and consuming the components in their product do we have instructions for installing and examples of importing and using a component and any other info they may need to get started consuming i m thinking we can follow
1
214,141
24,040,611,329
IssuesEvent
2022-09-16 01:06:43
LuisMartinSchick/website-portfolio
https://api.github.com/repos/LuisMartinSchick/website-portfolio
opened
WS-2022-0239 (Medium) detected in parse-url-6.0.0.tgz
security vulnerability
## WS-2022-0239 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-6.0.0.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz">https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz</a></p> <p> Dependency Hierarchy: - gatsby-2.32.13.tgz (Root Library) - gatsby-telemetry-1.10.2.tgz - git-up-4.0.5.tgz - :x: **parse-url-6.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Scripting via Improper Input Validation (parser differential) in parse-url before 8.0.0. Through this vulnerability, an attacker is capable to execute malicious JS codes. <p>Publish Date: 2022-07-02 <p>URL: <a href=https://github.com/ionicabizau/parse-url/commit/b88c81df8f4c5168af454eaa4f92afa9349e4e13>WS-2022-0239</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5fa3115f-5c97-4928-874c-3cc6302e154e">https://huntr.dev/bounties/5fa3115f-5c97-4928-874c-3cc6302e154e</a></p> <p>Release Date: 2022-07-02</p> <p>Fix Resolution (parse-url): 8.0.0</p> <p>Direct dependency fix Resolution (gatsby): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2022-0239 (Medium) detected in parse-url-6.0.0.tgz - ## WS-2022-0239 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-6.0.0.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz">https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz</a></p> <p> Dependency Hierarchy: - gatsby-2.32.13.tgz (Root Library) - gatsby-telemetry-1.10.2.tgz - git-up-4.0.5.tgz - :x: **parse-url-6.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Scripting via Improper Input Validation (parser differential) in parse-url before 8.0.0. Through this vulnerability, an attacker is capable to execute malicious JS codes. <p>Publish Date: 2022-07-02 <p>URL: <a href=https://github.com/ionicabizau/parse-url/commit/b88c81df8f4c5168af454eaa4f92afa9349e4e13>WS-2022-0239</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5fa3115f-5c97-4928-874c-3cc6302e154e">https://huntr.dev/bounties/5fa3115f-5c97-4928-874c-3cc6302e154e</a></p> <p>Release Date: 2022-07-02</p> <p>Fix Resolution (parse-url): 8.0.0</p> <p>Direct dependency fix Resolution (gatsby): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in parse url tgz ws medium severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href dependency hierarchy gatsby tgz root library gatsby telemetry tgz git up tgz x parse url tgz vulnerable library found in base branch master vulnerability details cross site scripting via improper input validation parser differential in parse url before through this vulnerability an attacker is capable to execute malicious js codes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url direct dependency fix resolution gatsby step up your open source security game with mend
0
18,110
24,136,145,710
IssuesEvent
2022-09-21 11:28:24
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Sign in screen > 'User not found' error message is getting displayed in the following scenario
Bug P1 Participant manager Process: Fixed
Steps: 1. Login to PM as Superadmin 2. Go to the Admins 3. Deactivate the active admin 4. Sign out 5. Try to sign in with the deactivated admin credential 6. Observe AR: 'User not found' error message is getting displayed ER: 'Your account has been deactivated' error message should get displayed in the above scenario
1.0
[PM] Sign in screen > 'User not found' error message is getting displayed in the following scenario - Steps: 1. Login to PM as Superadmin 2. Go to the Admins 3. Deactivate the active admin 4. Sign out 5. Try to sign in with the deactivated admin credential 6. Observe AR: 'User not found' error message is getting displayed ER: 'Your account has been deactivated' error message should get displayed in the above scenario
process
sign in screen user not found error message is getting displayed in the following scenario steps login to pm as superadmin go to the admins deactivate the active admin sign out try to sign in with the deactivated admin credential observe ar user not found error message is getting displayed er your account has been deactivated error message should get displayed in the above scenario
1
90,762
18,249,914,374
IssuesEvent
2021-10-02 03:01:32
firebase/firebase-ios-sdk
https://api.github.com/repos/firebase/firebase-ios-sdk
closed
Seeing potential issue in GoogleDataTransport that is blocking unit tests from running on Xcode 13
GoogleDataTransport Xcode 13 - iOS 15
<!-- DO NOT DELETE validate_template=true template_path=.github/ISSUE_TEMPLATE/bug_report.md --> ### Step 0: Are you in the right place? * For issues or feature requests related to __the code in this repository__ file a Github issue. * If this is a __feature request__ please use the Feature Request template. * For general technical questions, post a question on [StackOverflow](http://stackoverflow.com/) with the `firebase` tag. * For general (non-iOS) Firebase discussion, use the [firebase-talk](https://groups.google.com/forum/#!forum/firebase-talk) google group. * For backend issues, console issues, and other non-SDK help that does not fall under one of the above categories, reach out to [Firebase Support](https://firebase.google.com/support/). * Once you've read this section and determined that your issue is appropriate for this repository, please delete this section. ### [REQUIRED] Step 1: Describe your environment * Xcode version: __13___ * Firebase SDK version: __8.8___ * Installation method: `**CocoaPods** | Carthage | Zip file | Swift Package Manager` (select one) * Firebase Component: __GoogleDataTransport___ (Auth, Core, Database, Firestore, Messaging, Storage, etc) ### [REQUIRED] Step 2: Describe the problem #### Steps to reproduce: What happened? How can we make the problem occur? This could be a description, log/console output, etc. If you have a downloadable sample project that reproduces the bug you're reporting, you will likely receive a faster response on your issue. Several people on the internet/twitter are complaining that unit tests aren't running on Xcode 13. For others the solution was to update a 3rd party dependecy after reviewing the spindump file or taking samples from the simulator before the test runner times out. While I was reviewing the spindump file, in my case, it appears to be related to GoogleDataTransport 491 __52-[GDTCORUploadCoordinator uploadTargets:conditions:]_block_invoke + 535 (GDTCORUploadCoordinator.m:112,7 in GoogleDataTransport + 157303) [0x112324677] 1-491 I'm attaching a copy of the spindump for review. [Bootstrap Hang_0_65A9B5E4-AD68-453C-A455-C53F5E93CF7C.spindump.zip](https://github.com/firebase/firebase-ios-sdk/files/7255703/Bootstrap.Hang_0_65A9B5E4-AD68-453C-A455-C53F5E93CF7C.spindump.zip)
1.0
Seeing potential issue in GoogleDataTransport that is blocking unit tests from running on Xcode 13 - <!-- DO NOT DELETE validate_template=true template_path=.github/ISSUE_TEMPLATE/bug_report.md --> ### Step 0: Are you in the right place? * For issues or feature requests related to __the code in this repository__ file a Github issue. * If this is a __feature request__ please use the Feature Request template. * For general technical questions, post a question on [StackOverflow](http://stackoverflow.com/) with the `firebase` tag. * For general (non-iOS) Firebase discussion, use the [firebase-talk](https://groups.google.com/forum/#!forum/firebase-talk) google group. * For backend issues, console issues, and other non-SDK help that does not fall under one of the above categories, reach out to [Firebase Support](https://firebase.google.com/support/). * Once you've read this section and determined that your issue is appropriate for this repository, please delete this section. ### [REQUIRED] Step 1: Describe your environment * Xcode version: __13___ * Firebase SDK version: __8.8___ * Installation method: `**CocoaPods** | Carthage | Zip file | Swift Package Manager` (select one) * Firebase Component: __GoogleDataTransport___ (Auth, Core, Database, Firestore, Messaging, Storage, etc) ### [REQUIRED] Step 2: Describe the problem #### Steps to reproduce: What happened? How can we make the problem occur? This could be a description, log/console output, etc. If you have a downloadable sample project that reproduces the bug you're reporting, you will likely receive a faster response on your issue. Several people on the internet/twitter are complaining that unit tests aren't running on Xcode 13. For others the solution was to update a 3rd party dependecy after reviewing the spindump file or taking samples from the simulator before the test runner times out. While I was reviewing the spindump file, in my case, it appears to be related to GoogleDataTransport 491 __52-[GDTCORUploadCoordinator uploadTargets:conditions:]_block_invoke + 535 (GDTCORUploadCoordinator.m:112,7 in GoogleDataTransport + 157303) [0x112324677] 1-491 I'm attaching a copy of the spindump for review. [Bootstrap Hang_0_65A9B5E4-AD68-453C-A455-C53F5E93CF7C.spindump.zip](https://github.com/firebase/firebase-ios-sdk/files/7255703/Bootstrap.Hang_0_65A9B5E4-AD68-453C-A455-C53F5E93CF7C.spindump.zip)
non_process
seeing potential issue in googledatatransport that is blocking unit tests from running on xcode do not delete validate template true template path github issue template bug report md step are you in the right place for issues or feature requests related to the code in this repository file a github issue if this is a feature request please use the feature request template for general technical questions post a question on with the firebase tag for general non ios firebase discussion use the google group for backend issues console issues and other non sdk help that does not fall under one of the above categories reach out to once you ve read this section and determined that your issue is appropriate for this repository please delete this section step describe your environment xcode version firebase sdk version installation method cocoapods carthage zip file swift package manager select one firebase component googledatatransport auth core database firestore messaging storage etc step describe the problem steps to reproduce what happened how can we make the problem occur this could be a description log console output etc if you have a downloadable sample project that reproduces the bug you re reporting you will likely receive a faster response on your issue several people on the internet twitter are complaining that unit tests aren t running on xcode for others the solution was to update a party dependecy after reviewing the spindump file or taking samples from the simulator before the test runner times out while i was reviewing the spindump file in my case it appears to be related to googledatatransport block invoke gdtcoruploadcoordinator m in googledatatransport i m attaching a copy of the spindump for review
0
27,028
11,424,872,572
IssuesEvent
2020-02-03 18:39:37
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Is there a way to automate the on-boarding of a ASC Security Solution?
Pri3 cxp product-question security-center/svc triaged
Is there a way to automate the on-boarding of a ASC Security Solution? For Example, I would like to leverage Azure Policy to deploy the Qualys Security Solution across all of my subscriptions, with the the appropriate configuration, so that all VMs (existing and new) get teh Qualys Extension. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 747a2c52-c59c-dedc-9c72-6b6352ea2cab * Version Independent ID: 07e64d3e-26da-be39-de53-6e259c90545f * Content: [Built-in vulnerability scanner in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/built-in-vulnerability-assessment) * Content Source: [articles/security-center/built-in-vulnerability-assessment.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/built-in-vulnerability-assessment.md) * Service: **security-center** * GitHub Login: @memildin * Microsoft Alias: **memildin**
True
Is there a way to automate the on-boarding of a ASC Security Solution? - Is there a way to automate the on-boarding of a ASC Security Solution? For Example, I would like to leverage Azure Policy to deploy the Qualys Security Solution across all of my subscriptions, with the the appropriate configuration, so that all VMs (existing and new) get teh Qualys Extension. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 747a2c52-c59c-dedc-9c72-6b6352ea2cab * Version Independent ID: 07e64d3e-26da-be39-de53-6e259c90545f * Content: [Built-in vulnerability scanner in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/built-in-vulnerability-assessment) * Content Source: [articles/security-center/built-in-vulnerability-assessment.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/built-in-vulnerability-assessment.md) * Service: **security-center** * GitHub Login: @memildin * Microsoft Alias: **memildin**
non_process
is there a way to automate the on boarding of a asc security solution is there a way to automate the on boarding of a asc security solution for example i would like to leverage azure policy to deploy the qualys security solution across all of my subscriptions with the the appropriate configuration so that all vms existing and new get teh qualys extension document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id dedc version independent id content content source service security center github login memildin microsoft alias memildin
0
226,074
7,498,463,439
IssuesEvent
2018-04-09 04:48:55
ESAPI/esapi-java-legacy
https://api.github.com/repos/ESAPI/esapi-java-legacy
closed
ClassCastException when using ESAPI logger
Priority-Medium bug imported
_From [gautam...@gmail.com](https://code.google.com/u/110906976405804393731/) on May 17, 2013 03:57:18_ What steps will reproduce the problem? 1. Use a Maven project with Spring 3.0.3, ESAPI 2.0GA over Java 7 and Tomcat 7 2. Initialise logging CyborgSecurityConfiguration tmpConfig = new CyborgSecurityConfiguration((DefaultSecurityConfiguration) ESAPI.securityConfiguration()); ``` tmpConfig.setLogImplementation( "org.owasp.esapi.reference.Log4JLogFactory" ); tmpConfig.setLogLocation(logs.getAbsolutePath()+File.separator+"FastForward.log"); ESAPI.override(tmpConfig); log = (Log4JLogger) ESAPI.getLogger(LoggingManager.class); ``` 1. Initialise logger using Log4JLogger log = (Log4JLogger) ESAPI.getLogger(<classname>.class); in multiple classes What is the expected output? What do you see instead? Expected output is trace output. What is seen is java.lang.ClassCastException: org.apache.log4j.Logger cannot be cast to org.owasp.esapi.Logger at org.owasp.esapi.reference.Log4JLogFactory.getLogger(Log4JLogFactory.java:81) at org.owasp.esapi.ESAPI.getLogger(ESAPI.java:146) at com.cyborg.comm2.model.manager.MenuManager.<clinit>(MenuManager.java:28)… What version of the product are you using? On what operating system? ESAPI 2.0GA on Windows 7, Tomcat 7 and Java 7 Does this issue affect only a specified browser or set of browsers? All browsers, serverside issue Please provide any additional information below. _Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=299_
1.0
ClassCastException when using ESAPI logger - _From [gautam...@gmail.com](https://code.google.com/u/110906976405804393731/) on May 17, 2013 03:57:18_ What steps will reproduce the problem? 1. Use a Maven project with Spring 3.0.3, ESAPI 2.0GA over Java 7 and Tomcat 7 2. Initialise logging CyborgSecurityConfiguration tmpConfig = new CyborgSecurityConfiguration((DefaultSecurityConfiguration) ESAPI.securityConfiguration()); ``` tmpConfig.setLogImplementation( "org.owasp.esapi.reference.Log4JLogFactory" ); tmpConfig.setLogLocation(logs.getAbsolutePath()+File.separator+"FastForward.log"); ESAPI.override(tmpConfig); log = (Log4JLogger) ESAPI.getLogger(LoggingManager.class); ``` 1. Initialise logger using Log4JLogger log = (Log4JLogger) ESAPI.getLogger(<classname>.class); in multiple classes What is the expected output? What do you see instead? Expected output is trace output. What is seen is java.lang.ClassCastException: org.apache.log4j.Logger cannot be cast to org.owasp.esapi.Logger at org.owasp.esapi.reference.Log4JLogFactory.getLogger(Log4JLogFactory.java:81) at org.owasp.esapi.ESAPI.getLogger(ESAPI.java:146) at com.cyborg.comm2.model.manager.MenuManager.<clinit>(MenuManager.java:28)… What version of the product are you using? On what operating system? ESAPI 2.0GA on Windows 7, Tomcat 7 and Java 7 Does this issue affect only a specified browser or set of browsers? All browsers, serverside issue Please provide any additional information below. _Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=299_
non_process
classcastexception when using esapi logger from on may what steps will reproduce the problem use a maven project with spring esapi over java and tomcat initialise logging cyborgsecurityconfiguration tmpconfig new cyborgsecurityconfiguration defaultsecurityconfiguration esapi securityconfiguration tmpconfig setlogimplementation org owasp esapi reference tmpconfig setloglocation logs getabsolutepath file separator fastforward log esapi override tmpconfig log esapi getlogger loggingmanager class initialise logger using log esapi getlogger class in multiple classes what is the expected output what do you see instead expected output is trace output what is seen is java lang classcastexception org apache logger cannot be cast to org owasp esapi logger at org owasp esapi reference getlogger java at org owasp esapi esapi getlogger esapi java at com cyborg model manager menumanager menumanager java … what version of the product are you using on what operating system esapi on windows tomcat and java does this issue affect only a specified browser or set of browsers all browsers serverside issue please provide any additional information below original issue
0
246,853
20,921,235,145
IssuesEvent
2022-03-24 17:34:45
microsoft/playwright
https://api.github.com/repos/microsoft/playwright
closed
[Feature] Element screenshot crop option - stabilize visual regression by removing not important elements.
feature-visual-regression-testing v1.21
So I would like to have additional option in screenshot method which could add us possibility to crop a screenshot. ***Example*** - we are taking screenshot of the modal - button on the modal has spinning animations in the time of the loading - we don't care about button but we are interested in the content of the modal - problem is that our tool for screen comparison shows the difference between versions of the screenshots (master vs actual branch) because of the animation. So it would help other people to stabilize visual regression tests by removing not important elements from the screenshot. <img width="396" alt="image" src="https://user-images.githubusercontent.com/62831238/158187829-eeb2d82d-112c-4ff7-b83b-af376e3a2e3d.png"> It could be implemented similarly like test cafe did: https://testcafe.io/documentation/402676/reference/test-api/testcontroller/takeelementscreenshot#header It could look like that: ```await page.locator('.header').screenshot({ path: 'screenshot.png' , crop: {top: number, left: number, right: number, bottom: number}});``` Where crop takes coordinate associated within an element.
1.0
[Feature] Element screenshot crop option - stabilize visual regression by removing not important elements. - So I would like to have additional option in screenshot method which could add us possibility to crop a screenshot. ***Example*** - we are taking screenshot of the modal - button on the modal has spinning animations in the time of the loading - we don't care about button but we are interested in the content of the modal - problem is that our tool for screen comparison shows the difference between versions of the screenshots (master vs actual branch) because of the animation. So it would help other people to stabilize visual regression tests by removing not important elements from the screenshot. <img width="396" alt="image" src="https://user-images.githubusercontent.com/62831238/158187829-eeb2d82d-112c-4ff7-b83b-af376e3a2e3d.png"> It could be implemented similarly like test cafe did: https://testcafe.io/documentation/402676/reference/test-api/testcontroller/takeelementscreenshot#header It could look like that: ```await page.locator('.header').screenshot({ path: 'screenshot.png' , crop: {top: number, left: number, right: number, bottom: number}});``` Where crop takes coordinate associated within an element.
non_process
element screenshot crop option stabilize visual regression by removing not important elements so i would like to have additional option in screenshot method which could add us possibility to crop a screenshot example we are taking screenshot of the modal button on the modal has spinning animations in the time of the loading we don t care about button but we are interested in the content of the modal problem is that our tool for screen comparison shows the difference between versions of the screenshots master vs actual branch because of the animation so it would help other people to stabilize visual regression tests by removing not important elements from the screenshot img width alt image src it could be implemented similarly like test cafe did it could look like that await page locator header screenshot path screenshot png crop top number left number right number bottom number where crop takes coordinate associated within an element
0
357,211
10,603,589,438
IssuesEvent
2019-10-10 16:17:21
microsoft/terraform-provider-azuredevops
https://api.github.com/repos/microsoft/terraform-provider-azuredevops
closed
Faster builds on PR requests
priority-high
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The build that runs on every PR takes about 20 minutes currently because it tests on each OS and version of GO that are supported by the project. An alternative approach is to run this comprehensive build daily using a [scheduled job](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema#scheduled-trigger). Then the build for a PR can run with a single version of Go and a single OS, reducing the build time drastically.
1.0
Faster builds on PR requests - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The build that runs on every PR takes about 20 minutes currently because it tests on each OS and version of GO that are supported by the project. An alternative approach is to run this comprehensive build daily using a [scheduled job](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema#scheduled-trigger). Then the build for a PR can run with a single version of Go and a single OS, reducing the build time drastically.
non_process
faster builds on pr requests community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the build that runs on every pr takes about minutes currently because it tests on each os and version of go that are supported by the project an alternative approach is to run this comprehensive build daily using a then the build for a pr can run with a single version of go and a single os reducing the build time drastically
0
10,907
13,685,802,903
IssuesEvent
2020-09-30 07:44:55
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
opened
Increase code coverage
type: process
```>>nox -s cover``` ``` Name Stmts Miss Branch BrPart Cover Missing ---------------------------------------------------------------------------------- google/cloud/storage/__init__.py 8 0 0 0 100% google/cloud/storage/_helpers.py 110 0 27 0 100% google/cloud/storage/_http.py 15 0 2 0 100% google/cloud/storage/_signing.py 177 0 82 0 100% google/cloud/storage/acl.py 182 0 40 0 100% google/cloud/storage/batch.py 127 0 24 0 100% google/cloud/storage/blob.py 574 0 192 1 99% 818->816 google/cloud/storage/bucket.py 633 0 186 0 100% google/cloud/storage/client.py 211 0 68 0 100% google/cloud/storage/constants.py 13 0 0 0 100% google/cloud/storage/hmac_key.py 93 0 26 0 100% google/cloud/storage/iam.py 23 0 0 0 100% google/cloud/storage/notification.py 132 0 34 0 100% tests/unit/__init__.py 8 0 0 0 100% tests/unit/test__helpers.py 343 0 0 0 100% tests/unit/test__http.py 46 0 0 0 100% tests/unit/test__signing.py 504 0 58 0 100% tests/unit/test_acl.py 694 0 2 0 100% tests/unit/test_batch.py 392 0 6 0 100% tests/unit/test_blob.py 2693 0 162 0 100% tests/unit/test_bucket.py 2335 0 52 0 100% tests/unit/test_client.py 943 0 20 0 100% tests/unit/test_hmac_key.py 294 0 6 0 100% tests/unit/test_notification.py 313 0 2 0 100% ---------------------------------------------------------------------------------- TOTAL 10863 0 989 1 99% ```
1.0
Increase code coverage - ```>>nox -s cover``` ``` Name Stmts Miss Branch BrPart Cover Missing ---------------------------------------------------------------------------------- google/cloud/storage/__init__.py 8 0 0 0 100% google/cloud/storage/_helpers.py 110 0 27 0 100% google/cloud/storage/_http.py 15 0 2 0 100% google/cloud/storage/_signing.py 177 0 82 0 100% google/cloud/storage/acl.py 182 0 40 0 100% google/cloud/storage/batch.py 127 0 24 0 100% google/cloud/storage/blob.py 574 0 192 1 99% 818->816 google/cloud/storage/bucket.py 633 0 186 0 100% google/cloud/storage/client.py 211 0 68 0 100% google/cloud/storage/constants.py 13 0 0 0 100% google/cloud/storage/hmac_key.py 93 0 26 0 100% google/cloud/storage/iam.py 23 0 0 0 100% google/cloud/storage/notification.py 132 0 34 0 100% tests/unit/__init__.py 8 0 0 0 100% tests/unit/test__helpers.py 343 0 0 0 100% tests/unit/test__http.py 46 0 0 0 100% tests/unit/test__signing.py 504 0 58 0 100% tests/unit/test_acl.py 694 0 2 0 100% tests/unit/test_batch.py 392 0 6 0 100% tests/unit/test_blob.py 2693 0 162 0 100% tests/unit/test_bucket.py 2335 0 52 0 100% tests/unit/test_client.py 943 0 20 0 100% tests/unit/test_hmac_key.py 294 0 6 0 100% tests/unit/test_notification.py 313 0 2 0 100% ---------------------------------------------------------------------------------- TOTAL 10863 0 989 1 99% ```
process
increase code coverage nox s cover name stmts miss branch brpart cover missing google cloud storage init py google cloud storage helpers py google cloud storage http py google cloud storage signing py google cloud storage acl py google cloud storage batch py google cloud storage blob py google cloud storage bucket py google cloud storage client py google cloud storage constants py google cloud storage hmac key py google cloud storage iam py google cloud storage notification py tests unit init py tests unit test helpers py tests unit test http py tests unit test signing py tests unit test acl py tests unit test batch py tests unit test blob py tests unit test bucket py tests unit test client py tests unit test hmac key py tests unit test notification py total
1
19,806
26,188,343,671
IssuesEvent
2023-01-03 05:31:19
taikoxyz/taiko-mono
https://api.github.com/repos/taikoxyz/taiko-mono
closed
feat(bridge-ui): rework HORSE faucet
bridge feedback-processed
### Describe the feature request It's currently not user friendly to request HORSE. I'd suggest changing the icon so it looks less like a filter, changing the button text to 'Request HORSE' and align the button left. ### Describe alternatives you've considered Description of the alternatives you've considered here. ### Additional context ![Bridge Image](https://user-images.githubusercontent.com/60930264/209726045-7af551d7-e8b4-4c48-8a4c-2e0501d7cc1f.png)
1.0
feat(bridge-ui): rework HORSE faucet - ### Describe the feature request It's currently not user friendly to request HORSE. I'd suggest changing the icon so it looks less like a filter, changing the button text to 'Request HORSE' and align the button left. ### Describe alternatives you've considered Description of the alternatives you've considered here. ### Additional context ![Bridge Image](https://user-images.githubusercontent.com/60930264/209726045-7af551d7-e8b4-4c48-8a4c-2e0501d7cc1f.png)
process
feat bridge ui rework horse faucet describe the feature request it s currently not user friendly to request horse i d suggest changing the icon so it looks less like a filter changing the button text to request horse and align the button left describe alternatives you ve considered description of the alternatives you ve considered here additional context
1
3,179
6,255,346,466
IssuesEvent
2017-07-14 06:45:20
Pale-Moon-Addons-Team/phoebus
https://api.github.com/repos/Pale-Moon-Addons-Team/phoebus
closed
TRUNK should be refreshed
Application: PHOEBUS Component: AUS Component: DATABASE Component: SITE Module: funcProcessContent Status: FIXED
With all the edits being done on RELEASE, the TRUNK branch got out of sync and has old db data/code in it. It should be refreshed to reflect the current state of things by copying recent edits across to it from RELEASE.
1.0
TRUNK should be refreshed - With all the edits being done on RELEASE, the TRUNK branch got out of sync and has old db data/code in it. It should be refreshed to reflect the current state of things by copying recent edits across to it from RELEASE.
process
trunk should be refreshed with all the edits being done on release the trunk branch got out of sync and has old db data code in it it should be refreshed to reflect the current state of things by copying recent edits across to it from release
1
263,554
28,040,507,210
IssuesEvent
2023-03-28 18:07:54
socialtables/react-image-fallback
https://api.github.com/repos/socialtables/react-image-fallback
closed
CVE-2020-7793 (High) detected in ua-parser-js-0.7.18.tgz - autoclosed
security vulnerability
## CVE-2020-7793 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.18.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.18.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.18.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - prop-types-15.6.1.tgz (Root Library) - fbjs-0.8.16.tgz - :x: **ua-parser-js-0.7.18.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/socialtables/react-image-fallback/commit/c7153a7f2cc175073dc8b6baf6898558c21c66a0">c7153a7f2cc175073dc8b6baf6898558c21c66a0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info). <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution (ua-parser-js): 0.7.23</p> <p>Direct dependency fix Resolution (prop-types): 15.6.2</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prop-types","packageVersion":"15.6.1","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"prop-types:15.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"15.6.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7793","vulnerabilityDetails":"The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7793 (High) detected in ua-parser-js-0.7.18.tgz - autoclosed - ## CVE-2020-7793 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.18.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.18.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.18.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - prop-types-15.6.1.tgz (Root Library) - fbjs-0.8.16.tgz - :x: **ua-parser-js-0.7.18.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/socialtables/react-image-fallback/commit/c7153a7f2cc175073dc8b6baf6898558c21c66a0">c7153a7f2cc175073dc8b6baf6898558c21c66a0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info). <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution (ua-parser-js): 0.7.23</p> <p>Direct dependency fix Resolution (prop-types): 15.6.2</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prop-types","packageVersion":"15.6.1","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"prop-types:15.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"15.6.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7793","vulnerabilityDetails":"The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in ua parser js tgz autoclosed cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file package json path to vulnerable library node modules ua parser js package json dependency hierarchy prop types tgz root library fbjs tgz x ua parser js tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos in multiple regexes see linked commit for more info publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ua parser js direct dependency fix resolution prop types rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree prop types isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the package ua parser js before are vulnerable to regular expression denial of service redos in multiple regexes see linked commit for more info vulnerabilityurl
0
345,438
24,859,330,773
IssuesEvent
2022-10-27 06:54:05
POSTECH-CVLab/PeRFception
https://api.github.com/repos/POSTECH-CVLab/PeRFception
closed
Downloading PeRFception-ScanNet
documentation
I believe the CO3d dataset was moved from one-drive to hugging face, however, I am unable to find a working download link for the PeRFceptionScanNet. Can you provide a new download link? thank you in advance Additionally what was the estimated time to produce the PeRFception-ScanNet given your setup. Thank you again
1.0
Downloading PeRFception-ScanNet - I believe the CO3d dataset was moved from one-drive to hugging face, however, I am unable to find a working download link for the PeRFceptionScanNet. Can you provide a new download link? thank you in advance Additionally what was the estimated time to produce the PeRFception-ScanNet given your setup. Thank you again
non_process
downloading perfception scannet i believe the dataset was moved from one drive to hugging face however i am unable to find a working download link for the perfceptionscannet can you provide a new download link thank you in advance additionally what was the estimated time to produce the perfception scannet given your setup thank you again
0
215,442
16,603,419,767
IssuesEvent
2021-06-01 23:08:06
elyra-ai/elyra
https://api.github.com/repos/elyra-ai/elyra
closed
install doc (et al): replace all mentions of `pip` with `pip3`
area:documentation good first issue kind:enhancement program:Jumpstart program:RCOS
**Is your feature request related to a problem? Please describe.** The `pip` command isn't necessarily pointing to a python 3 binary. To avoid any issues, we should update the documentation to use `pip3` instead of `pip` in example commands: - `README.md` - `developer_guide/development_workflow.md` - `getting_started/installation.md` - `recipes/creating-a-custom-runtime-image.md` **Describe the solution you'd like** A clear and concise description of what you want to happen. `s/pip/pip3/g`
1.0
install doc (et al): replace all mentions of `pip` with `pip3` - **Is your feature request related to a problem? Please describe.** The `pip` command isn't necessarily pointing to a python 3 binary. To avoid any issues, we should update the documentation to use `pip3` instead of `pip` in example commands: - `README.md` - `developer_guide/development_workflow.md` - `getting_started/installation.md` - `recipes/creating-a-custom-runtime-image.md` **Describe the solution you'd like** A clear and concise description of what you want to happen. `s/pip/pip3/g`
non_process
install doc et al replace all mentions of pip with is your feature request related to a problem please describe the pip command isn t necessarily pointing to a python binary to avoid any issues we should update the documentation to use instead of pip in example commands readme md developer guide development workflow md getting started installation md recipes creating a custom runtime image md describe the solution you d like a clear and concise description of what you want to happen s pip g
0
51,913
27,300,180,931
IssuesEvent
2023-02-24 00:48:47
moby/moby
https://api.github.com/repos/moby/moby
closed
Memory usage of containerd-shim
area/runtime area/performance
This is a question and concern. With every container a containerd-shim process is started. It seems to be that that process consumes between 3-5mb (excluding shared memory). Doing some quick, not so scientific tests, I have seen the memory usage of my system take ~400mb more than 1.10.3 to launch 100 nginx containers. This was a quick cursory test, so I wanted to ask the maintainers how they envision this working going forward, before I dug in much more doing performance tests on 1.11. I also have a concern about VM overcommit because each containerd-shim process consumes over 200mb of virtual memory. I honestly don't see how this approach will scale with this intermediate process. With the current usage pattern if I run 500 containers (which is a very common use case), I will be using 2gb of memory just for Docker. Am I missing something obvious here? This doesn't seem good.
True
Memory usage of containerd-shim - This is a question and concern. With every container a containerd-shim process is started. It seems to be that that process consumes between 3-5mb (excluding shared memory). Doing some quick, not so scientific tests, I have seen the memory usage of my system take ~400mb more than 1.10.3 to launch 100 nginx containers. This was a quick cursory test, so I wanted to ask the maintainers how they envision this working going forward, before I dug in much more doing performance tests on 1.11. I also have a concern about VM overcommit because each containerd-shim process consumes over 200mb of virtual memory. I honestly don't see how this approach will scale with this intermediate process. With the current usage pattern if I run 500 containers (which is a very common use case), I will be using 2gb of memory just for Docker. Am I missing something obvious here? This doesn't seem good.
non_process
memory usage of containerd shim this is a question and concern with every container a containerd shim process is started it seems to be that that process consumes between excluding shared memory doing some quick not so scientific tests i have seen the memory usage of my system take more than to launch nginx containers this was a quick cursory test so i wanted to ask the maintainers how they envision this working going forward before i dug in much more doing performance tests on i also have a concern about vm overcommit because each containerd shim process consumes over of virtual memory i honestly don t see how this approach will scale with this intermediate process with the current usage pattern if i run containers which is a very common use case i will be using of memory just for docker am i missing something obvious here this doesn t seem good
0
145,399
5,575,088,508
IssuesEvent
2017-03-28 00:27:07
autoboxer/MARE
https://api.github.com/repos/autoboxer/MARE
closed
gallery search not showing results
beta launch bug node priority: high
no matter how inclusive the search terms, results are never shown.
1.0
gallery search not showing results - no matter how inclusive the search terms, results are never shown.
non_process
gallery search not showing results no matter how inclusive the search terms results are never shown
0
183,743
14,248,321,847
IssuesEvent
2020-11-19 12:46:44
kaspanet/kaspad
https://api.github.com/repos/kaspanet/kaspad
opened
Add Windows to the CI and make sure the tests pass also on windows
CI tests
We might want to use this opportunity to migrate away from jenkins.
1.0
Add Windows to the CI and make sure the tests pass also on windows - We might want to use this opportunity to migrate away from jenkins.
non_process
add windows to the ci and make sure the tests pass also on windows we might want to use this opportunity to migrate away from jenkins
0
13,807
16,567,035,416
IssuesEvent
2021-05-29 16:25:29
SmartHalo-User-Group/SmartHalo-2
https://api.github.com/repos/SmartHalo-User-Group/SmartHalo-2
closed
Takes too long for SH Support to respond to non-shipping related issues
process improvement needed
Long delays before Support responds to messages sent to support@smarthalo.bike (> 1-2 weeks). This is seriously harming the SmartHalo reputation. Recommendation to split ticket stream: 1) delivery related stream 2) dead-on-arrival / damaged during shipping 3) functional / technical issues ad 1) Create a page where backers/buyers can see the shipping status (per country/state/batch) ad 2) scan the queue daily for anything that seems to be related to units that arrived dead-on-arrival and/or have been damaged during shipping. This needs to be addressed with the highest possible priority. ad 3) group the functional/technical issues and make use of the issue tracker that the SmartHalo User Group is maintaining. Things escalate if users are not seeing that their issues are being responded to. If a personal message is (temporarily) not possible, then a publically accessible issue tracker is the next best thing.
1.0
Takes too long for SH Support to respond to non-shipping related issues - Long delays before Support responds to messages sent to support@smarthalo.bike (> 1-2 weeks). This is seriously harming the SmartHalo reputation. Recommendation to split ticket stream: 1) delivery related stream 2) dead-on-arrival / damaged during shipping 3) functional / technical issues ad 1) Create a page where backers/buyers can see the shipping status (per country/state/batch) ad 2) scan the queue daily for anything that seems to be related to units that arrived dead-on-arrival and/or have been damaged during shipping. This needs to be addressed with the highest possible priority. ad 3) group the functional/technical issues and make use of the issue tracker that the SmartHalo User Group is maintaining. Things escalate if users are not seeing that their issues are being responded to. If a personal message is (temporarily) not possible, then a publically accessible issue tracker is the next best thing.
process
takes too long for sh support to respond to non shipping related issues long delays before support responds to messages sent to support smarthalo bike weeks this is seriously harming the smarthalo reputation recommendation to split ticket stream delivery related stream dead on arrival damaged during shipping functional technical issues ad create a page where backers buyers can see the shipping status per country state batch ad scan the queue daily for anything that seems to be related to units that arrived dead on arrival and or have been damaged during shipping this needs to be addressed with the highest possible priority ad group the functional technical issues and make use of the issue tracker that the smarthalo user group is maintaining things escalate if users are not seeing that their issues are being responded to if a personal message is temporarily not possible then a publically accessible issue tracker is the next best thing
1
21,391
29,202,231,994
IssuesEvent
2023-05-21 00:37:40
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Híbrido | São Paulo/SP, Rio de Janeiro/RJ e Belo Horizonte/MG] Analistas DevOps - Nível Sênior | HrSoul
TELECOM PYTHON DOCKER KUBERNETES DEVOPS AWS LINUX PROCESSOS SHELL JENKINS CI CD UMA ECS ANSIBLE TERRAFORM HELP WANTED HTTP IAC EKS YAML Stale
Olá! Sou o Lucas Oliveira, trabalho com Recrutamento e Seleção - TI, na HrSoul Consultoria. Sobre a HrSoul: Criada visando atender as necessidades do mercado de TI/Telecom e Serviços, suprindo a carência de consultorias com expertise na área, assim como inovar os processos de Recursos Humanos para os demais segmentos. ![HrSoul Logo](https://user-images.githubusercontent.com/98987631/212729834-9b4138f0-83ac-4cf3-8876-e9af159185ba.png) Oferecemos oportunidade para profissionais com o seguinte perfil: ## **Analistas DevOps - Nível Sênior (3 vagas) - SP/RJ/MG** Sólida experiência com Throubleshooting em ambientes com IaC e CI/CD utilizando Python e YAML; Experiencia com criação de scripts (shell, bash Linux e Python); Vivência com AWS, microsserviços, Nexus, EKS, Lambda, ECS; Bons conhecimentos em Terraform, Cloud Formation, Ansible, Jenkins e CodeDeploy; Conhecimentos em Orquestração de Docker e Kubernetes; Imprescindível disponibilidade para atuação em modelo híbrido em uma das 3 regiões: São Paulo, Rio de Janeiro e/ou Belo Horizonte. Mais informações podem ser alinhadas conosco durante o processo, você conhece alguém que possa se interessar? Indicações para Aileen Azevedo - [aazevedo@hrsoul.com.br](mailto:aazevedo@hrsoul.com.br) ou cadastre seu CV em nosso site – [www.hrsoul.com.br](http://www.hrsoul.com.br/) Muito obrigado!
1.0
[Híbrido | São Paulo/SP, Rio de Janeiro/RJ e Belo Horizonte/MG] Analistas DevOps - Nível Sênior | HrSoul - Olá! Sou o Lucas Oliveira, trabalho com Recrutamento e Seleção - TI, na HrSoul Consultoria. Sobre a HrSoul: Criada visando atender as necessidades do mercado de TI/Telecom e Serviços, suprindo a carência de consultorias com expertise na área, assim como inovar os processos de Recursos Humanos para os demais segmentos. ![HrSoul Logo](https://user-images.githubusercontent.com/98987631/212729834-9b4138f0-83ac-4cf3-8876-e9af159185ba.png) Oferecemos oportunidade para profissionais com o seguinte perfil: ## **Analistas DevOps - Nível Sênior (3 vagas) - SP/RJ/MG** Sólida experiência com Throubleshooting em ambientes com IaC e CI/CD utilizando Python e YAML; Experiencia com criação de scripts (shell, bash Linux e Python); Vivência com AWS, microsserviços, Nexus, EKS, Lambda, ECS; Bons conhecimentos em Terraform, Cloud Formation, Ansible, Jenkins e CodeDeploy; Conhecimentos em Orquestração de Docker e Kubernetes; Imprescindível disponibilidade para atuação em modelo híbrido em uma das 3 regiões: São Paulo, Rio de Janeiro e/ou Belo Horizonte. Mais informações podem ser alinhadas conosco durante o processo, você conhece alguém que possa se interessar? Indicações para Aileen Azevedo - [aazevedo@hrsoul.com.br](mailto:aazevedo@hrsoul.com.br) ou cadastre seu CV em nosso site – [www.hrsoul.com.br](http://www.hrsoul.com.br/) Muito obrigado!
process
analistas devops nível sênior hrsoul olá sou o lucas oliveira trabalho com recrutamento e seleção ti na hrsoul consultoria sobre a hrsoul criada visando atender as necessidades do mercado de ti telecom e serviços suprindo a carência de consultorias com expertise na área assim como inovar os processos de recursos humanos para os demais segmentos oferecemos oportunidade para profissionais com o seguinte perfil analistas devops nível sênior vagas sp rj mg sólida experiência com throubleshooting em ambientes com iac e ci cd utilizando python e yaml experiencia com criação de scripts shell bash linux e python vivência com aws microsserviços nexus eks lambda ecs bons conhecimentos em terraform cloud formation ansible jenkins e codedeploy conhecimentos em orquestração de docker e kubernetes imprescindível disponibilidade para atuação em modelo híbrido em uma das regiões são paulo rio de janeiro e ou belo horizonte mais informações podem ser alinhadas conosco durante o processo você conhece alguém que possa se interessar indicações para aileen azevedo mailto aazevedo hrsoul com br ou cadastre seu cv em nosso site – muito obrigado
1