Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
66,620
8,955,416,885
IssuesEvent
2019-01-26 07:16:52
memezinga/memezinga-legacy
https://api.github.com/repos/memezinga/memezinga-legacy
closed
Release v0.2.0 Pinhead
Documentation
### v0.2.0 Pinhead ![pinhead](https://static.comicvine.com/uploads/original/12/123441/3246521-pinhead_wallpaper_by_evilflesh.jpg) ### Marketing/Social - [ ] #10 Liar a la gente y birras pagadas ### Features - [ ] #80 Implementación de Router y Plantillas - [ ] #67 Marca de agua - [ ] #47 Login con Github usando Firebase - [ ] #51 Perfil del usuario - [ ] #49 Gestión de multimedia - [ ] #50 Link del Meme generado - [ ] #45 Compartir en redes sociales... - [ ] #46 Migración de los templates de MEMES - [ ] #44 Lazy loading para los memes de la galeria - [ ] #48 Newsletter con Mailchimp ### Bugs - [ ] #42 Refactorizar Javascript - [ ] #78 Code style... ### Review - [ ] Calidad del código - [ ] Test :trollface:
1.0
Release v0.2.0 Pinhead - ### v0.2.0 Pinhead ![pinhead](https://static.comicvine.com/uploads/original/12/123441/3246521-pinhead_wallpaper_by_evilflesh.jpg) ### Marketing/Social - [ ] #10 Liar a la gente y birras pagadas ### Features - [ ] #80 Implementación de Router y Plantillas - [ ] #67 Marca de agua - [ ] #47 Login con Github usando Firebase - [ ] #51 Perfil del usuario - [ ] #49 Gestión de multimedia - [ ] #50 Link del Meme generado - [ ] #45 Compartir en redes sociales... - [ ] #46 Migración de los templates de MEMES - [ ] #44 Lazy loading para los memes de la galeria - [ ] #48 Newsletter con Mailchimp ### Bugs - [ ] #42 Refactorizar Javascript - [ ] #78 Code style... ### Review - [ ] Calidad del código - [ ] Test :trollface:
non_defect
release pinhead pinhead marketing social liar a la gente y birras pagadas features implementación de router y plantillas marca de agua login con github usando firebase perfil del usuario gestión de multimedia link del meme generado compartir en redes sociales migración de los templates de memes lazy loading para los memes de la galeria newsletter con mailchimp bugs refactorizar javascript code style review calidad del código test trollface
0
53,396
13,261,517,209
IssuesEvent
2020-08-20 20:02:32
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
paraboloid - coverage of private interfaces is weak (Trac #1305)
Migrated from Trac combo reconstruction defect
a serialization and a "known failure" test will quickly bump coverage <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1305">https://code.icecube.wisc.edu/projects/icecube/ticket/1305</a>, reported by negaand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2016-02-10T10:59:05", "_ts": "1455101945934200", "description": "a serialization and a \"known failure\" test will quickly bump coverage", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2015-08-28T23:07:43", "component": "combo reconstruction", "summary": "paraboloid - coverage of private interfaces is weak", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
paraboloid - coverage of private interfaces is weak (Trac #1305) - a serialization and a "known failure" test will quickly bump coverage <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1305">https://code.icecube.wisc.edu/projects/icecube/ticket/1305</a>, reported by negaand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2016-02-10T10:59:05", "_ts": "1455101945934200", "description": "a serialization and a \"known failure\" test will quickly bump coverage", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2015-08-28T23:07:43", "component": "combo reconstruction", "summary": "paraboloid - coverage of private interfaces is weak", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
paraboloid coverage of private interfaces is weak trac a serialization and a known failure test will quickly bump coverage migrated from json status closed changetime ts description a serialization and a known failure test will quickly bump coverage reporter nega cc resolution fixed time component combo reconstruction summary paraboloid coverage of private interfaces is weak priority normal keywords milestone owner kjmeagher type defect
1
262,647
8,272,290,910
IssuesEvent
2018-09-16 18:35:05
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
Move preInvoke/postInvoke logic from TransactionManagerHelper to JavaEETransactionManagerSimplified to avoid different code paths for lookup vs injection
Component: jts ERR: Assignee Priority: Major Type: Improvement ee7ri_cleanup_deferred
TransactionManagerHelper shouldn't check for the servlet invocation, and be responsible for enlist/delist resources. This might also remove the need for the EJB container to call special preInvoke/postInvoke methods, but do just suspend/resume calls. #### Affected Versions [V3]
1.0
Move preInvoke/postInvoke logic from TransactionManagerHelper to JavaEETransactionManagerSimplified to avoid different code paths for lookup vs injection - TransactionManagerHelper shouldn't check for the servlet invocation, and be responsible for enlist/delist resources. This might also remove the need for the EJB container to call special preInvoke/postInvoke methods, but do just suspend/resume calls. #### Affected Versions [V3]
non_defect
move preinvoke postinvoke logic from transactionmanagerhelper to javaeetransactionmanagersimplified to avoid different code paths for lookup vs injection transactionmanagerhelper shouldn t check for the servlet invocation and be responsible for enlist delist resources this might also remove the need for the ejb container to call special preinvoke postinvoke methods but do just suspend resume calls affected versions
0
155,658
5,958,840,233
IssuesEvent
2017-05-29 09:11:42
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
Negative usage on asds
priority_normal type_bug
``` alba version 0.9.5 git_revision: "tags/0.9.5-0-ge4cf366" git_repo: "https://github.com/openvstorage/alba.git" compile_time: "11/05/2016 15:40:26 UTC" machine: "455badec0dde 3.13.0-24-generic x86_64 x86_64 x86_64 GNU/Linux" compiler_version: "4.02.3" ``` On regular time intervals, asds/backend usage shows negative values. (see screenshots) After restarting the asd the negative value disappears and the correct value is shown. Negative used_bytes: ``` root@str02:~# alba list-osds --config etcd://127.0.0.1:2379/ovs/arakoon/be-backend-abm/config | grep oyR 2016/05/26 15:20:10 286478 +0200 - str02 - 1403/0 - alba/cli - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/arakoon/be-backend-abm/config "oyRgZLZKDUVXts1birJ1IceDxUJ32EsH"); other = "{ \"id\" : \"oyRgZLZKDUVXts1birJ1IceDxUJ32EsH\", \"node_id\" : \"qNqsb2zoMLDE7iCAkSBpNz24a27iwDNX\", \"port\" : 8603, \"used_bytes\" : \"-187152090328\", \"total_bytes\" : \"1999018582016\", \"version\" : \"AsdV1\", \"network_interfaces\":[ ]}"; ``` Restarting asd (correct value): ``` root@str02:~# restart alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH start/running, process 10051 root@str02:~# alba list-osds --config etcd://127.0.0.1:2379/ovs/arakoon/be-backend-abm/config | grep oyR 2016/05/26 15:23:05 632690 +0200 - str02 - 11000/0 - alba/cli - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/arakoon/be-backend-abm/config "oyRgZLZKDUVXts1birJ1IceDxUJ32EsH"); other = "{ \"id\" : \"oyRgZLZKDUVXts1birJ1IceDxUJ32EsH\", \"node_id\" : \"qNqsb2zoMLDE7iCAkSBpNz24a27iwDNX\", \"port\" : 8603, \"used_bytes\" : \"874780016\", \"total_bytes\" : \"1999018582016\", \"version\" : \"AsdV1\", \"network_interfaces\":[ ]}"; ``` ![backend_size](https://cloud.githubusercontent.com/assets/2900612/15576398/f2e89a2c-2357-11e6-95e7-d359c2cf84ea.png) ![negative_asd](https://cloud.githubusercontent.com/assets/2900612/15576397/f2e32308-2357-11e6-9a40-0183b1eda3be.png) [alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH.txt](https://github.com/openvstorage/alba/files/284467/alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH.txt)
1.0
Negative usage on asds - ``` alba version 0.9.5 git_revision: "tags/0.9.5-0-ge4cf366" git_repo: "https://github.com/openvstorage/alba.git" compile_time: "11/05/2016 15:40:26 UTC" machine: "455badec0dde 3.13.0-24-generic x86_64 x86_64 x86_64 GNU/Linux" compiler_version: "4.02.3" ``` On regular time intervals, asds/backend usage shows negative values. (see screenshots) After restarting the asd the negative value disappears and the correct value is shown. Negative used_bytes: ``` root@str02:~# alba list-osds --config etcd://127.0.0.1:2379/ovs/arakoon/be-backend-abm/config | grep oyR 2016/05/26 15:20:10 286478 +0200 - str02 - 1403/0 - alba/cli - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/arakoon/be-backend-abm/config "oyRgZLZKDUVXts1birJ1IceDxUJ32EsH"); other = "{ \"id\" : \"oyRgZLZKDUVXts1birJ1IceDxUJ32EsH\", \"node_id\" : \"qNqsb2zoMLDE7iCAkSBpNz24a27iwDNX\", \"port\" : 8603, \"used_bytes\" : \"-187152090328\", \"total_bytes\" : \"1999018582016\", \"version\" : \"AsdV1\", \"network_interfaces\":[ ]}"; ``` Restarting asd (correct value): ``` root@str02:~# restart alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH start/running, process 10051 root@str02:~# alba list-osds --config etcd://127.0.0.1:2379/ovs/arakoon/be-backend-abm/config | grep oyR 2016/05/26 15:23:05 632690 +0200 - str02 - 11000/0 - alba/cli - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/arakoon/be-backend-abm/config "oyRgZLZKDUVXts1birJ1IceDxUJ32EsH"); other = "{ \"id\" : \"oyRgZLZKDUVXts1birJ1IceDxUJ32EsH\", \"node_id\" : \"qNqsb2zoMLDE7iCAkSBpNz24a27iwDNX\", \"port\" : 8603, \"used_bytes\" : \"874780016\", \"total_bytes\" : \"1999018582016\", \"version\" : \"AsdV1\", \"network_interfaces\":[ ]}"; ``` ![backend_size](https://cloud.githubusercontent.com/assets/2900612/15576398/f2e89a2c-2357-11e6-95e7-d359c2cf84ea.png) ![negative_asd](https://cloud.githubusercontent.com/assets/2900612/15576397/f2e32308-2357-11e6-9a40-0183b1eda3be.png) [alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH.txt](https://github.com/openvstorage/alba/files/284467/alba-asd-oyRgZLZKDUVXts1birJ1IceDxUJ32EsH.txt)
non_defect
negative usage on asds alba version git revision tags git repo compile time utc machine generic gnu linux compiler version on regular time intervals asds backend usage shows negative values see screenshots after restarting the asd the negative value disappears and the correct value is shown negative used bytes root alba list osds config etcd ovs arakoon be backend abm config grep oyr alba cli info etcd etcdctl peers get ovs arakoon be backend abm config other id node id port used bytes total bytes version network interfaces restarting asd correct value root restart alba asd alba asd start running process root alba list osds config etcd ovs arakoon be backend abm config grep oyr alba cli info etcd etcdctl peers get ovs arakoon be backend abm config other id node id port used bytes total bytes version network interfaces
0
16,390
2,891,245,802
IssuesEvent
2015-06-15 02:24:13
ibus/ibus
https://api.github.com/repos/ibus/ibus
closed
Radio menu items does not work on IBus panel icon with GTK 3.16
Component-ibus Priority-Medium Type-Defect
``` GTK 3.16 has cleared the radio buttons' state in gtk_radio_menu_item_set_group(): https://git.gnome.org/browse/gtk+/commit/?id=955aed92272b88a72617cbadca6f3010920a4f1f IBus sets the state before set_group() is called so the radio menu items does not work, e.g. ibus-anthy Input Mode menu items. The check menu items no longer work as radio menu item. To reproduce: 1. Activate an engine which has radio menus or check menus likes ibus-anthy in ibus gtk panel with GTK 3.16. 2. See the radio menus. All item states are cleared and cannot select any. ``` Original issue reported on code.google.com by `takao.fujiwara1` on 2015-05-22 02:57:29
1.0
Radio menu items does not work on IBus panel icon with GTK 3.16 - ``` GTK 3.16 has cleared the radio buttons' state in gtk_radio_menu_item_set_group(): https://git.gnome.org/browse/gtk+/commit/?id=955aed92272b88a72617cbadca6f3010920a4f1f IBus sets the state before set_group() is called so the radio menu items does not work, e.g. ibus-anthy Input Mode menu items. The check menu items no longer work as radio menu item. To reproduce: 1. Activate an engine which has radio menus or check menus likes ibus-anthy in ibus gtk panel with GTK 3.16. 2. See the radio menus. All item states are cleared and cannot select any. ``` Original issue reported on code.google.com by `takao.fujiwara1` on 2015-05-22 02:57:29
defect
radio menu items does not work on ibus panel icon with gtk gtk has cleared the radio buttons state in gtk radio menu item set group ibus sets the state before set group is called so the radio menu items does not work e g ibus anthy input mode menu items the check menu items no longer work as radio menu item to reproduce activate an engine which has radio menus or check menus likes ibus anthy in ibus gtk panel with gtk see the radio menus all item states are cleared and cannot select any original issue reported on code google com by takao on
1
72,029
23,900,018,919
IssuesEvent
2022-09-08 17:53:02
zorzella/guiceberry
https://api.github.com/repos/zorzella/guiceberry
closed
Replace tl4j with guava-testlib
Priority-Medium Type-Defect auto-migrated
``` guiceberry depends on the unsupported tl4j library. guiceberry should use guava-testlib, its successor, instead of tl4j. The only complication would be the GuiceBerryJunit3TestCase. However, almost no one uses junit 3 anymore, so deleting it would be a good way to get rid of tl4j. That package doesn't work for the majority of users today anyway because tl4j is not in a Maven repository, so anyone using guiceberry from Maven will experience runtime errors until they manually include tl4j in their projects. At the very least we should deprecate GuiceBerryJunit3TestCase, but junit 3 itself has been being phased out for so long that I doubt it really matters. ``` Original issue reported on code.google.com by `benjamin...@gmail.com` on 5 Jun 2013 at 7:35
1.0
Replace tl4j with guava-testlib - ``` guiceberry depends on the unsupported tl4j library. guiceberry should use guava-testlib, its successor, instead of tl4j. The only complication would be the GuiceBerryJunit3TestCase. However, almost no one uses junit 3 anymore, so deleting it would be a good way to get rid of tl4j. That package doesn't work for the majority of users today anyway because tl4j is not in a Maven repository, so anyone using guiceberry from Maven will experience runtime errors until they manually include tl4j in their projects. At the very least we should deprecate GuiceBerryJunit3TestCase, but junit 3 itself has been being phased out for so long that I doubt it really matters. ``` Original issue reported on code.google.com by `benjamin...@gmail.com` on 5 Jun 2013 at 7:35
defect
replace with guava testlib guiceberry depends on the unsupported library guiceberry should use guava testlib its successor instead of the only complication would be the however almost no one uses junit anymore so deleting it would be a good way to get rid of that package doesn t work for the majority of users today anyway because is not in a maven repository so anyone using guiceberry from maven will experience runtime errors until they manually include in their projects at the very least we should deprecate but junit itself has been being phased out for so long that i doubt it really matters original issue reported on code google com by benjamin gmail com on jun at
1
68,837
14,958,803,395
IssuesEvent
2021-01-27 01:37:48
mihorsky/phpbb
https://api.github.com/repos/mihorsky/phpbb
opened
CVE-2019-18797 (Medium) detected in opennmsopennms-source-26.0.0-1
security vulnerability
## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-26.0.0-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/mihorsky/phpbb/commit/df86840378ba9a67de78b42213b0c85da2b3eadc">df86840378ba9a67de78b42213b0c85da2b3eadc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>phpbb/node_modules/node-sass/src/libsass/src/eval.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: LibSass - 3.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-18797 (Medium) detected in opennmsopennms-source-26.0.0-1 - ## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-26.0.0-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/mihorsky/phpbb/commit/df86840378ba9a67de78b42213b0c85da2b3eadc">df86840378ba9a67de78b42213b0c85da2b3eadc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>phpbb/node_modules/node-sass/src/libsass/src/eval.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: LibSass - 3.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in opennmsopennms source cve medium severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href found in base branch master vulnerable source files phpbb node modules node sass src libsass src eval cpp vulnerability details libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
15,847
28,106,846,780
IssuesEvent
2023-03-31 01:57:31
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Renovate sidecar for yarn does not pass npm config to child process
type:bug reproduction:needed status:requirements priority-5-triage stale
### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 34.48.4-slim ### If you're self-hosting Renovate, select which platform you are using. GitLab self-hosted ### If you're self-hosting Renovate, tell us what version of the platform you run. 15.8.3 ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug ## General config The most important setting is that our renovateconfig allows us to authenticate against an artifactory instance. If you want something detailed i can provide it ## Given the following files in our repo .npmrc: ``` @myscopedpackage:registry=https://artifactory.example.com/artifactory/api/npm/myrepo/ ``` renovate.json ```json { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "dependencyDashboard": true, "lockFileMaintenance": { "enabled": true, "schedule": [ "at any time" ] } } ``` In the errors below we can see that the npm-configs are not passed to the sidecar processes it seems. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: lock file error (repository=myrepo/renovate-test, branch=renovate/test) "err": { "name": "ExecError", "cmd": "/bin/sh -c yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts", "stderr": "error Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n", "stdout": "yarn install v1.22.19\n[1/4] Resolving packages...\ninfo Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command.\n", "options": { "cwd": "/tmp/renovate/repos/gitlab/myrepo/renovate-test", "encoding": "utf-8", "env": { "NPM_CONFIG_CACHE": "/tmp/renovate/cache/others/npm", "npm_config_store": "/tmp/renovate/cache/others/pnpm", "CI": "true", "YARN_CACHE_FOLDER": "/tmp/renovate/cache/others/yarn", "HOME": "/home/ubuntu", "PATH": "/home/ubuntu/.cargo/bin:/home/ubuntu/.local/bin:/go/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.11.1/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "LC_ALL": "C.UTF-8", "LANG": "C.UTF-8", "BUILDPACK_CACHE_DIR": "/tmp/renovate/cache/containerbase", "CONTAINERBASE_CACHE_DIR": "/tmp/renovate/cache/containerbase" }, "maxBuffer": 10485760, "timeout": 900000 }, "exitCode": 1, "message": "Command failed: yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts\nerror Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n", "stack": "ExecError: Command failed: yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts\nerror Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n\n at ChildProcess.<anonymous> (/usr/src/app/node_modules/renovate/lib/util/exec/common.ts:99:11)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)" }, "type": "yarn" ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
1.0
Renovate sidecar for yarn does not pass npm config to child process - ### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 34.48.4-slim ### If you're self-hosting Renovate, select which platform you are using. GitLab self-hosted ### If you're self-hosting Renovate, tell us what version of the platform you run. 15.8.3 ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug ## General config The most important setting is that our renovateconfig allows us to authenticate against an artifactory instance. If you want something detailed i can provide it ## Given the following files in our repo .npmrc: ``` @myscopedpackage:registry=https://artifactory.example.com/artifactory/api/npm/myrepo/ ``` renovate.json ```json { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "dependencyDashboard": true, "lockFileMaintenance": { "enabled": true, "schedule": [ "at any time" ] } } ``` In the errors below we can see that the npm-configs are not passed to the sidecar processes it seems. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: lock file error (repository=myrepo/renovate-test, branch=renovate/test) "err": { "name": "ExecError", "cmd": "/bin/sh -c yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts", "stderr": "error Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n", "stdout": "yarn install v1.22.19\n[1/4] Resolving packages...\ninfo Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command.\n", "options": { "cwd": "/tmp/renovate/repos/gitlab/myrepo/renovate-test", "encoding": "utf-8", "env": { "NPM_CONFIG_CACHE": "/tmp/renovate/cache/others/npm", "npm_config_store": "/tmp/renovate/cache/others/pnpm", "CI": "true", "YARN_CACHE_FOLDER": "/tmp/renovate/cache/others/yarn", "HOME": "/home/ubuntu", "PATH": "/home/ubuntu/.cargo/bin:/home/ubuntu/.local/bin:/go/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.11.1/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "LC_ALL": "C.UTF-8", "LANG": "C.UTF-8", "BUILDPACK_CACHE_DIR": "/tmp/renovate/cache/containerbase", "CONTAINERBASE_CACHE_DIR": "/tmp/renovate/cache/containerbase" }, "maxBuffer": 10485760, "timeout": 900000 }, "exitCode": 1, "message": "Command failed: yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts\nerror Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n", "stack": "ExecError: Command failed: yarn install --ignore-engines --ignore-platform --network-timeout 100000 --ignore-scripts\nerror Couldn't find package \"@myscopedpackage/mysdk\" on the \"npm\" registry.\n\n at ChildProcess.<anonymous> (/usr/src/app/node_modules/renovate/lib/util/exec/common.ts:99:11)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)" }, "type": "yarn" ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
non_defect
renovate sidecar for yarn does not pass npm config to child process how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run slim if you re self hosting renovate select which platform you are using gitlab self hosted if you re self hosting renovate tell us what version of the platform you run was this something which used to work for you and then stopped i never saw this working describe the bug general config the most important setting is that our renovateconfig allows us to authenticate against an artifactory instance if you want something detailed i can provide it given the following files in our repo npmrc myscopedpackage registry renovate json json schema dependencydashboard true lockfilemaintenance enabled true schedule at any time in the errors below we can see that the npm configs are not passed to the sidecar processes it seems relevant debug logs logs debug lock file error repository myrepo renovate test branch renovate test err name execerror cmd bin sh c yarn install ignore engines ignore platform network timeout ignore scripts stderr error couldn t find package myscopedpackage mysdk on the npm registry n stdout yarn install n resolving packages ninfo visit for documentation about this command n options cwd tmp renovate repos gitlab myrepo renovate test encoding utf env npm config cache tmp renovate cache others npm npm config store tmp renovate cache others pnpm ci true yarn cache folder tmp renovate cache others yarn home home ubuntu path home ubuntu cargo bin home ubuntu local bin go bin home ubuntu bin opt buildpack tools python bin home ubuntu npm global bin home ubuntu bin usr local sbin usr local bin usr sbin usr bin sbin bin lc all c utf lang c utf buildpack cache dir tmp renovate cache containerbase containerbase cache dir tmp renovate cache containerbase maxbuffer timeout exitcode message command failed yarn install ignore engines ignore platform network timeout ignore scripts nerror couldn t find package myscopedpackage mysdk on the npm registry n stack execerror command failed yarn install ignore engines ignore platform network timeout ignore scripts nerror couldn t find package myscopedpackage mysdk on the npm registry n n at childprocess usr src app node modules renovate lib util exec common ts n at childprocess emit node events n at childprocess emit node domain n at process childprocess handle onexit node internal child process type yarn have you created a minimal reproduction repository no reproduction repository
0
57,407
24,120,433,968
IssuesEvent
2022-09-20 18:11:44
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
vets-api k8s db migrations
console-services
## Description This ticket provides details that we will need to add initContainers to the specification to run database migrations for the vets-api k8s offering. Currently, the k8s offering AND the current BRD/EC2 offering both use the same db instance. We will want to think about adding the db migration task to initContainers once we fully switch over to the k8s offering. We need to run migrations as part of the k8s deployment process to ensure that the db is up to date. ## Tasks - [x] Discover any overlap between the BRD and K8s deployments - [x] Compare the to Databases - [x] Write the k8s specifications for the migrations ### Other notes: - Discussed during the meeting on 6/10, we may be able to run migrations via a sidecar strategy. - ArgoCD uses Presync jobs to run schema migrations. This replaces the need to use initContainers. https://argo-cd.readthedocs.io/en/stable/user-guide/resource_hooks/ - Helm charts use pre-install jobs to run commands on pods before new application version is install in the new pods. https://helm.sh/docs/topics/charts_hooks/ - Migrations are rarely needed. It the past, there have been only around 2 migrations per month. Throughout AB testing, when a migration needs to be ran, the PR will be merged during low traffic times. This will allow the migration to be ran concurrently in both setups to prevent any migration errors. ## Definition of Done - [x] Define an initiContainers or a preSync job spec to run migrations in the k8s deployment manifest - [PreSync Hook PR](https://github.com/department-of-veterans-affairs/vsp-infra-application-manifests/pull/792) - [ ] Create Github tag to prevent migrations from merging into master until traffic is at an all time low. - [x] Verify that migrations run successfully in k8s
1.0
vets-api k8s db migrations - ## Description This ticket provides details that we will need to add initContainers to the specification to run database migrations for the vets-api k8s offering. Currently, the k8s offering AND the current BRD/EC2 offering both use the same db instance. We will want to think about adding the db migration task to initContainers once we fully switch over to the k8s offering. We need to run migrations as part of the k8s deployment process to ensure that the db is up to date. ## Tasks - [x] Discover any overlap between the BRD and K8s deployments - [x] Compare the to Databases - [x] Write the k8s specifications for the migrations ### Other notes: - Discussed during the meeting on 6/10, we may be able to run migrations via a sidecar strategy. - ArgoCD uses Presync jobs to run schema migrations. This replaces the need to use initContainers. https://argo-cd.readthedocs.io/en/stable/user-guide/resource_hooks/ - Helm charts use pre-install jobs to run commands on pods before new application version is install in the new pods. https://helm.sh/docs/topics/charts_hooks/ - Migrations are rarely needed. It the past, there have been only around 2 migrations per month. Throughout AB testing, when a migration needs to be ran, the PR will be merged during low traffic times. This will allow the migration to be ran concurrently in both setups to prevent any migration errors. ## Definition of Done - [x] Define an initiContainers or a preSync job spec to run migrations in the k8s deployment manifest - [PreSync Hook PR](https://github.com/department-of-veterans-affairs/vsp-infra-application-manifests/pull/792) - [ ] Create Github tag to prevent migrations from merging into master until traffic is at an all time low. - [x] Verify that migrations run successfully in k8s
non_defect
vets api db migrations description this ticket provides details that we will need to add initcontainers to the specification to run database migrations for the vets api offering currently the offering and the current brd offering both use the same db instance we will want to think about adding the db migration task to initcontainers once we fully switch over to the offering we need to run migrations as part of the deployment process to ensure that the db is up to date tasks discover any overlap between the brd and deployments compare the to databases write the specifications for the migrations other notes discussed during the meeting on we may be able to run migrations via a sidecar strategy argocd uses presync jobs to run schema migrations this replaces the need to use initcontainers helm charts use pre install jobs to run commands on pods before new application version is install in the new pods migrations are rarely needed it the past there have been only around migrations per month throughout ab testing when a migration needs to be ran the pr will be merged during low traffic times this will allow the migration to be ran concurrently in both setups to prevent any migration errors definition of done define an initicontainers or a presync job spec to run migrations in the deployment manifest create github tag to prevent migrations from merging into master until traffic is at an all time low verify that migrations run successfully in
0
34,659
14,481,679,725
IssuesEvent
2020-12-10 12:58:55
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Support for Azure Media Services Configuration Options
enhancement services/media
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description I'm relatively new to Terraform. I have seen the resource for azurerm_media_services_account but i cannot find any examples of how to configure more options within. For example how do you connect up a CDN profile/endpoint? Setup Live Streaming options. Is there an example for this? Or is there another way of doing this without resorting to an arm template resource? ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> azurerm_media_services_account ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> https://www.terraform.io/docs/providers/azurerm/r/media_services_account.html
1.0
Support for Azure Media Services Configuration Options - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description I'm relatively new to Terraform. I have seen the resource for azurerm_media_services_account but i cannot find any examples of how to configure more options within. For example how do you connect up a CDN profile/endpoint? Setup Live Streaming options. Is there an example for this? Or is there another way of doing this without resorting to an arm template resource? ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> azurerm_media_services_account ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> https://www.terraform.io/docs/providers/azurerm/r/media_services_account.html
non_defect
support for azure media services configuration options community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description i m relatively new to terraform i have seen the resource for azurerm media services account but i cannot find any examples of how to configure more options within for example how do you connect up a cdn profile endpoint setup live streaming options is there an example for this or is there another way of doing this without resorting to an arm template resource new or affected resource s azurerm media services account references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
226,066
17,946,869,930
IssuesEvent
2021-09-12 00:42:20
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Back-end Developer @TOTVS
CLT JavaScript TDD Java Remoto DevOps TypeScript Spring Testes Unitários Angular Git HTML CSS Stale
## Nossa empresa Somos a maior empresa de tecnologia do Brasil e nosso poder de transformação faz com que nossos softwares estejam presentes em mais de 40mil empresas no Brasil e no mundo. O nosso grande segredo para estar sempre um passo a frente do mercado é a ousadia e a transformação que nossos TOTVERS promovem todos os dias. Ousadia para movimentar 25% do PIB brasileiro por meio dos clientes que usam os nossos softwares. Ousadia para ser a primeira empresa de tecnologia do Brasil a ser listada na Ibovespa. Ousadia para fazer da mudança a nossa rotina, com mais de 40 aquisições ao longo de nossa história. Nossa curiosidade nos impulsiona e, o melhor, todo mundo colabora com todo mundo, não importa quem deu a ideia: todos estão dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados. Tudo isso acontece em um ambiente leve e divertido, mas já deixamos claro: a intensidade e a rapidez com que as coisas acontecem aqui é acima da média. ## Descrição da vaga Desenvolverá soluções baseadas em microsserviços/cloud; Utilizará técnicas como TDD e DDD e metodologia Ágil; Desenvolverá soluções técnicas de alta complexidade foco na usabilidade do produto; Garantirá a utilização e aplicação de boas práticas no desenvolvimento de testes; Fará estudos com clientes e áreas envolvidas no projeto; Contribuirá com a disseminação do conhecimento adquiridos sobre técnicas e tecnologias; Irá propor inovações que se apliquem ao roadmap; Participará de discussões táticas e de definição de metas da área; Orientará e capacitará o time de desenvolvimento para aplicação das soluções arquitetadas. ## Local Remoto ## Requisitos **Obrigatórios:** JAVA Nível avançado Spring Nível avançado Domínio em arquitetura web; Testes unitários e end to end (e2e) - nível avançado; Métodos ágeis de desenvolvimento de software – nível intermediário; Sistemas distribuídos e escalabilidade – nível intermediário. **Desejáveis:** Software de edição de imagens e prototipação de interface – nível básico; Formação superior na área de TI; Angular Nível avançado; HTML CSS Javascript/Typescript. Performance de Front-end– nível avançado; Integração e Entrega contínua de Front-end – nível avançado; Conhecimento do mercado de logística; Conhecimento prático de modelagem DDD; Conhecimento prático de DevOps; Projetos disponíveis para avaliação no Git. ## Benefícios Plano de saúde Plano odontológico Vale refeição/alimentação Licença maternidade e paternidade estendida Seguro de vida Previdência privada Gympass Universidade corporativa com conteúdo e certificações gratuitos para os nossos TOTVERS Escritório que estimula a criatividade e produtividade com ambientes para lanches, salas de jogos, mesas de bilhar e poltronas para relaxar Cuidamos dos nossos TOTVERS com o Programa +Saudáveis, com assessoria e ações voltadas para o bem estar em corpo, mente e finanças pessoais Programa +Vantagens, a maior rede de descontos da América Latina, exclusivos para nossos TOTVERS Valorizamos a diversidade e inclusão, consideramos todos os candidatos para as nossas vagas independente de cor, raça, religião, gênero e identidade de gênero, nacionalidade, deficiência, orientação sexual, ascendência, idade, etc ## Contratação CLT ## Como se candidatar Por favor cadastre-se no link: https://vagas.hunterco.com.br/job/60c299acf79840001d26ae33?c=4145 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 03 dias após cada processo. E-mail para contato em caso de não haver resposta: janaina.nessi@parceirohunterco.com.br ## Labels #### Alocação - Remoto #### Regime - CLT #### Nível - Sênior
1.0
[Remoto] Back-end Developer @TOTVS - ## Nossa empresa Somos a maior empresa de tecnologia do Brasil e nosso poder de transformação faz com que nossos softwares estejam presentes em mais de 40mil empresas no Brasil e no mundo. O nosso grande segredo para estar sempre um passo a frente do mercado é a ousadia e a transformação que nossos TOTVERS promovem todos os dias. Ousadia para movimentar 25% do PIB brasileiro por meio dos clientes que usam os nossos softwares. Ousadia para ser a primeira empresa de tecnologia do Brasil a ser listada na Ibovespa. Ousadia para fazer da mudança a nossa rotina, com mais de 40 aquisições ao longo de nossa história. Nossa curiosidade nos impulsiona e, o melhor, todo mundo colabora com todo mundo, não importa quem deu a ideia: todos estão dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados. Tudo isso acontece em um ambiente leve e divertido, mas já deixamos claro: a intensidade e a rapidez com que as coisas acontecem aqui é acima da média. ## Descrição da vaga Desenvolverá soluções baseadas em microsserviços/cloud; Utilizará técnicas como TDD e DDD e metodologia Ágil; Desenvolverá soluções técnicas de alta complexidade foco na usabilidade do produto; Garantirá a utilização e aplicação de boas práticas no desenvolvimento de testes; Fará estudos com clientes e áreas envolvidas no projeto; Contribuirá com a disseminação do conhecimento adquiridos sobre técnicas e tecnologias; Irá propor inovações que se apliquem ao roadmap; Participará de discussões táticas e de definição de metas da área; Orientará e capacitará o time de desenvolvimento para aplicação das soluções arquitetadas. ## Local Remoto ## Requisitos **Obrigatórios:** JAVA Nível avançado Spring Nível avançado Domínio em arquitetura web; Testes unitários e end to end (e2e) - nível avançado; Métodos ágeis de desenvolvimento de software – nível intermediário; Sistemas distribuídos e escalabilidade – nível intermediário. **Desejáveis:** Software de edição de imagens e prototipação de interface – nível básico; Formação superior na área de TI; Angular Nível avançado; HTML CSS Javascript/Typescript. Performance de Front-end– nível avançado; Integração e Entrega contínua de Front-end – nível avançado; Conhecimento do mercado de logística; Conhecimento prático de modelagem DDD; Conhecimento prático de DevOps; Projetos disponíveis para avaliação no Git. ## Benefícios Plano de saúde Plano odontológico Vale refeição/alimentação Licença maternidade e paternidade estendida Seguro de vida Previdência privada Gympass Universidade corporativa com conteúdo e certificações gratuitos para os nossos TOTVERS Escritório que estimula a criatividade e produtividade com ambientes para lanches, salas de jogos, mesas de bilhar e poltronas para relaxar Cuidamos dos nossos TOTVERS com o Programa +Saudáveis, com assessoria e ações voltadas para o bem estar em corpo, mente e finanças pessoais Programa +Vantagens, a maior rede de descontos da América Latina, exclusivos para nossos TOTVERS Valorizamos a diversidade e inclusão, consideramos todos os candidatos para as nossas vagas independente de cor, raça, religião, gênero e identidade de gênero, nacionalidade, deficiência, orientação sexual, ascendência, idade, etc ## Contratação CLT ## Como se candidatar Por favor cadastre-se no link: https://vagas.hunterco.com.br/job/60c299acf79840001d26ae33?c=4145 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 03 dias após cada processo. E-mail para contato em caso de não haver resposta: janaina.nessi@parceirohunterco.com.br ## Labels #### Alocação - Remoto #### Regime - CLT #### Nível - Sênior
non_defect
back end developer totvs nossa empresa somos a maior empresa de tecnologia do brasil e nosso poder de transformação faz com que nossos softwares estejam presentes em mais de empresas no brasil e no mundo o nosso grande segredo para estar sempre um passo a frente do mercado é a ousadia e a transformação que nossos totvers promovem todos os dias ousadia para movimentar do pib brasileiro por meio dos clientes que usam os nossos softwares ousadia para ser a primeira empresa de tecnologia do brasil a ser listada na ibovespa ousadia para fazer da mudança a nossa rotina com mais de aquisições ao longo de nossa história nossa curiosidade nos impulsiona e o melhor todo mundo colabora com todo mundo não importa quem deu a ideia todos estão dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados tudo isso acontece em um ambiente leve e divertido mas já deixamos claro a intensidade e a rapidez com que as coisas acontecem aqui é acima da média descrição da vaga desenvolverá soluções baseadas em microsserviços cloud utilizará técnicas como tdd e ddd e metodologia ágil desenvolverá soluções técnicas de alta complexidade foco na usabilidade do produto garantirá a utilização e aplicação de boas práticas no desenvolvimento de testes fará estudos com clientes e áreas envolvidas no projeto contribuirá com a disseminação do conhecimento adquiridos sobre técnicas e tecnologias irá propor inovações que se apliquem ao roadmap participará de discussões táticas e de definição de metas da área orientará e capacitará o time de desenvolvimento para aplicação das soluções arquitetadas local remoto requisitos obrigatórios java nível avançado spring nível avançado domínio em arquitetura web testes unitários e end to end nível avançado métodos ágeis de desenvolvimento de software – nível intermediário sistemas distribuídos e escalabilidade – nível intermediário desejáveis software de edição de imagens e prototipação de interface – nível básico formação superior na área de ti angular nível avançado html css javascript typescript performance de front end– nível avançado integração e entrega contínua de front end – nível avançado conhecimento do mercado de logística conhecimento prático de modelagem ddd conhecimento prático de devops projetos disponíveis para avaliação no git benefícios plano de saúde plano odontológico vale refeição alimentação licença maternidade e paternidade estendida seguro de vida previdência privada gympass universidade corporativa com conteúdo e certificações gratuitos para os nossos totvers escritório que estimula a criatividade e produtividade com ambientes para lanches salas de jogos mesas de bilhar e poltronas para relaxar cuidamos dos nossos totvers com o programa saudáveis com assessoria e ações voltadas para o bem estar em corpo mente e finanças pessoais programa vantagens a maior rede de descontos da américa latina exclusivos para nossos totvers valorizamos a diversidade e inclusão consideramos todos os candidatos para as nossas vagas independente de cor raça religião gênero e identidade de gênero nacionalidade deficiência orientação sexual ascendência idade etc contratação clt como se candidatar por favor cadastre se no link tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta janaina nessi parceirohunterco com br labels alocação remoto regime clt nível sênior
0
27,210
4,931,762,825
IssuesEvent
2016-11-28 11:17:09
johnny-y-wang/hash-identifier
https://api.github.com/repos/johnny-y-wang/hash-identifier
closed
Python 3.3 compatibility errors
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Run the script in Python 3.3 What is the expected output? What do you see instead? compilation errors ( print is replace with print(). and raw_input() with input().) What version of the product are you using? On what operating system? Python 3.3, win7, hash identifier 1.1 Please provide any additional information below. Please find attached the corrected script to run with Python 3.3, not backward compatatible http://www.python.org/dev/peps/pep-3105/#backwards-compatibility ``` Original issue reported on code.google.com by `michail....@gmail.com` on 12 Nov 2012 at 9:48 Attachments: - [Hash_ID_v2.0.py](https://storage.googleapis.com/google-code-attachments/hash-identifier/issue-1/comment-0/Hash_ID_v2.0.py)
1.0
Python 3.3 compatibility errors - ``` What steps will reproduce the problem? 1. Run the script in Python 3.3 What is the expected output? What do you see instead? compilation errors ( print is replace with print(). and raw_input() with input().) What version of the product are you using? On what operating system? Python 3.3, win7, hash identifier 1.1 Please provide any additional information below. Please find attached the corrected script to run with Python 3.3, not backward compatatible http://www.python.org/dev/peps/pep-3105/#backwards-compatibility ``` Original issue reported on code.google.com by `michail....@gmail.com` on 12 Nov 2012 at 9:48 Attachments: - [Hash_ID_v2.0.py](https://storage.googleapis.com/google-code-attachments/hash-identifier/issue-1/comment-0/Hash_ID_v2.0.py)
defect
python compatibility errors what steps will reproduce the problem run the script in python what is the expected output what do you see instead compilation errors print is replace with print and raw input with input what version of the product are you using on what operating system python hash identifier please provide any additional information below please find attached the corrected script to run with python not backward compatatible original issue reported on code google com by michail gmail com on nov at attachments
1
11,421
2,651,391,384
IssuesEvent
2015-03-16 11:07:57
douglasdrumond/cloaked-computing-machine
https://api.github.com/repos/douglasdrumond/cloaked-computing-machine
opened
[CLOSED] Pressing the <insert> key displays a questionmark instead of going to insert mode
auto-migrated Priority-Medium Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Mar 16, 2015 at 09:15 GMT_ _Originally opened as https://github.com/douglasdrumond/macvim/issues/9_ ---- ``` In command mode, pressing <insert> (on an usb keyboard) switches the mouse cursor to a "?". * In command mode I expect to enter insert mode. * In insert mode I expect to enter replace mode. * In replace mode I expect to enter insert mode. ``` Original issue reported on code.google.com by `christ...@zagrodnick.de` on 24 Aug 2007 at 2:05
1.0
[CLOSED] Pressing the <insert> key displays a questionmark instead of going to insert mode - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Mar 16, 2015 at 09:15 GMT_ _Originally opened as https://github.com/douglasdrumond/macvim/issues/9_ ---- ``` In command mode, pressing <insert> (on an usb keyboard) switches the mouse cursor to a "?". * In command mode I expect to enter insert mode. * In insert mode I expect to enter replace mode. * In replace mode I expect to enter insert mode. ``` Original issue reported on code.google.com by `christ...@zagrodnick.de` on 24 Aug 2007 at 2:05
defect
pressing the key displays a questionmark instead of going to insert mode issue by monday mar at gmt originally opened as in command mode pressing on an usb keyboard switches the mouse cursor to a in command mode i expect to enter insert mode in insert mode i expect to enter replace mode in replace mode i expect to enter insert mode original issue reported on code google com by christ zagrodnick de on aug at
1
16,614
2,920,434,510
IssuesEvent
2015-06-24 18:55:09
ashanbh/chrome-rest-client
https://api.github.com/repos/ashanbh/chrome-rest-client
closed
Content-Type
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Call REST API POST Method 2. Add Headers Content-Type: application/json 3. also select Content-Type from drop down "application/json" A) What is the expected output? What do you see instead? When i call the API in the request header i am seeing Content-Type: multipart/form-data; boundary=----WebKitFormBoundarysB8WKHBozuM7UQVY But It should be application/json. B) On what operating system, browser and browser version? I am suing client as Windows 7 with Chrome browser 31.x, and my REST API running in REDHAT C) Please provide any additional information below. Is there any setting i need to do to get the Content-Type as application/json" over writing multipart/form ``` Original issue reported on code.google.com by `ramesh.c...@gmail.com` on 15 Nov 2013 at 11:08
1.0
Content-Type - ``` What steps will reproduce the problem? 1. Call REST API POST Method 2. Add Headers Content-Type: application/json 3. also select Content-Type from drop down "application/json" A) What is the expected output? What do you see instead? When i call the API in the request header i am seeing Content-Type: multipart/form-data; boundary=----WebKitFormBoundarysB8WKHBozuM7UQVY But It should be application/json. B) On what operating system, browser and browser version? I am suing client as Windows 7 with Chrome browser 31.x, and my REST API running in REDHAT C) Please provide any additional information below. Is there any setting i need to do to get the Content-Type as application/json" over writing multipart/form ``` Original issue reported on code.google.com by `ramesh.c...@gmail.com` on 15 Nov 2013 at 11:08
defect
content type what steps will reproduce the problem call rest api post method add headers content type application json also select content type from drop down application json a what is the expected output what do you see instead when i call the api in the request header i am seeing content type multipart form data boundary but it should be application json b on what operating system browser and browser version i am suing client as windows with chrome browser x and my rest api running in redhat c please provide any additional information below is there any setting i need to do to get the content type as application json over writing multipart form original issue reported on code google com by ramesh c gmail com on nov at
1
229,281
18,287,866,666
IssuesEvent
2021-10-05 12:23:21
chipsalliance/Surelog
https://api.github.com/repos/chipsalliance/Surelog
closed
Associative array parsed as int
sv-tests
In the following file, the associative array `a` is parsed as `int_var` and element selection is parsed as `bit_select`. ```systemverilog module top; int a [int]; assign a[5] = 5; endmodule ``` Here is uhdm: ``` slpp_all/surelog.uhdm: Restored design Pre-Elab: design: (work@top) |vpiName:work@top |uhdmallModules: \_module: work@top (work@top) associative_array.sv:1:1: , endln:4:10, parent:work@top |vpiDefName:work@top |vpiFullName:work@top |vpiContAssign: \_cont_assign: , line:3:11, endln:3:19, parent:work@top |vpiRhs: \_constant: , line:3:18, endln:3:19 |vpiConstType:9 |vpiDecompile:5 |vpiSize:64 |UINT:5 |vpiLhs: \_bit_select: (work@top.a), line:3:11, endln:3:12, parent:work@top.a |vpiName:a |vpiFullName:work@top.a |vpiIndex: \_constant: , line:3:13, endln:3:14, parent:work@top.a |vpiConstType:9 |vpiDecompile:5 |vpiSize:64 |UINT:5 |vpiNet: \_logic_net: (work@top.a), line:2:8, parent:work@top |vpiName:a |vpiFullName:work@top.a |uhdmtopModules: \_module: work@top (work@top) associative_array.sv:1:1: , endln:4:10 |vpiDefName:work@top |vpiName:work@top |vpiVariables: \_array_var: (work@top.a), line:2:8, endln:2:9, parent:work@top |vpiArrayType:1 |vpiName:a |vpiFullName:work@top.a |vpiAutomatic:1 |vpiRandType:1 |vpiVisibility:1 |vpiSize:1 |vpiReg: \_int_var: (work@top.a), parent:work@top.a |vpiFullName:work@top.a ```
1.0
Associative array parsed as int - In the following file, the associative array `a` is parsed as `int_var` and element selection is parsed as `bit_select`. ```systemverilog module top; int a [int]; assign a[5] = 5; endmodule ``` Here is uhdm: ``` slpp_all/surelog.uhdm: Restored design Pre-Elab: design: (work@top) |vpiName:work@top |uhdmallModules: \_module: work@top (work@top) associative_array.sv:1:1: , endln:4:10, parent:work@top |vpiDefName:work@top |vpiFullName:work@top |vpiContAssign: \_cont_assign: , line:3:11, endln:3:19, parent:work@top |vpiRhs: \_constant: , line:3:18, endln:3:19 |vpiConstType:9 |vpiDecompile:5 |vpiSize:64 |UINT:5 |vpiLhs: \_bit_select: (work@top.a), line:3:11, endln:3:12, parent:work@top.a |vpiName:a |vpiFullName:work@top.a |vpiIndex: \_constant: , line:3:13, endln:3:14, parent:work@top.a |vpiConstType:9 |vpiDecompile:5 |vpiSize:64 |UINT:5 |vpiNet: \_logic_net: (work@top.a), line:2:8, parent:work@top |vpiName:a |vpiFullName:work@top.a |uhdmtopModules: \_module: work@top (work@top) associative_array.sv:1:1: , endln:4:10 |vpiDefName:work@top |vpiName:work@top |vpiVariables: \_array_var: (work@top.a), line:2:8, endln:2:9, parent:work@top |vpiArrayType:1 |vpiName:a |vpiFullName:work@top.a |vpiAutomatic:1 |vpiRandType:1 |vpiVisibility:1 |vpiSize:1 |vpiReg: \_int_var: (work@top.a), parent:work@top.a |vpiFullName:work@top.a ```
non_defect
associative array parsed as int in the following file the associative array a is parsed as int var and element selection is parsed as bit select systemverilog module top int a assign a endmodule here is uhdm slpp all surelog uhdm restored design pre elab design work top vpiname work top uhdmallmodules module work top work top associative array sv endln parent work top vpidefname work top vpifullname work top vpicontassign cont assign line endln parent work top vpirhs constant line endln vpiconsttype vpidecompile vpisize uint vpilhs bit select work top a line endln parent work top a vpiname a vpifullname work top a vpiindex constant line endln parent work top a vpiconsttype vpidecompile vpisize uint vpinet logic net work top a line parent work top vpiname a vpifullname work top a uhdmtopmodules module work top work top associative array sv endln vpidefname work top vpiname work top vpivariables array var work top a line endln parent work top vpiarraytype vpiname a vpifullname work top a vpiautomatic vpirandtype vpivisibility vpisize vpireg int var work top a parent work top a vpifullname work top a
0
134,061
18,411,794,252
IssuesEvent
2021-10-13 06:58:41
ChoeMinji/mongo-r4.4.6
https://api.github.com/repos/ChoeMinji/mongo-r4.4.6
opened
CVE-2018-1000872 (Medium) detected in PyKMIP-0.4.0.tar.gz
security vulnerability
## CVE-2018-1000872 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>PyKMIP-0.4.0.tar.gz</b></p></summary> <p>KMIP library</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/f6/204b675c5a6d4cf0efa01734728bbb443d784cf9ed510597dc94bb2ba26b/PyKMIP-0.4.0.tar.gz">https://files.pythonhosted.org/packages/3d/f6/204b675c5a6d4cf0efa01734728bbb443d784cf9ed510597dc94bb2ba26b/PyKMIP-0.4.0.tar.gz</a></p> <p>Path to dependency file: mongo-r4.4.6/src/third_party/wiredtiger/lang/python</p> <p>Path to vulnerable library: /src/third_party/wiredtiger/lang/python,/src/third_party/wiredtiger/bench/workgen</p> <p> Dependency Hierarchy: - :x: **PyKMIP-0.4.0.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/mongo-r4.4.6/commit/9c4537f1af3987a4f237e73712977c87c207c818">9c4537f1af3987a4f237e73712977c87c207c818</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OpenKMIP PyKMIP version All versions before 0.8.0 contains a CWE 399: Resource Management Errors (similar issue to CVE-2015-5262) vulnerability in PyKMIP server that can result in DOS: the server can be made unavailable by one or more clients opening all of the available sockets. This attack appear to be exploitable via A client or clients open sockets with the server and then never close them. This vulnerability appears to have been fixed in 0.8.0. <p>Publish Date: 2018-12-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000872>CVE-2018-1000872</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000872">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000872</a></p> <p>Release Date: 2018-12-20</p> <p>Fix Resolution: v0.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1000872 (Medium) detected in PyKMIP-0.4.0.tar.gz - ## CVE-2018-1000872 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>PyKMIP-0.4.0.tar.gz</b></p></summary> <p>KMIP library</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/f6/204b675c5a6d4cf0efa01734728bbb443d784cf9ed510597dc94bb2ba26b/PyKMIP-0.4.0.tar.gz">https://files.pythonhosted.org/packages/3d/f6/204b675c5a6d4cf0efa01734728bbb443d784cf9ed510597dc94bb2ba26b/PyKMIP-0.4.0.tar.gz</a></p> <p>Path to dependency file: mongo-r4.4.6/src/third_party/wiredtiger/lang/python</p> <p>Path to vulnerable library: /src/third_party/wiredtiger/lang/python,/src/third_party/wiredtiger/bench/workgen</p> <p> Dependency Hierarchy: - :x: **PyKMIP-0.4.0.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/mongo-r4.4.6/commit/9c4537f1af3987a4f237e73712977c87c207c818">9c4537f1af3987a4f237e73712977c87c207c818</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OpenKMIP PyKMIP version All versions before 0.8.0 contains a CWE 399: Resource Management Errors (similar issue to CVE-2015-5262) vulnerability in PyKMIP server that can result in DOS: the server can be made unavailable by one or more clients opening all of the available sockets. This attack appear to be exploitable via A client or clients open sockets with the server and then never close them. This vulnerability appears to have been fixed in 0.8.0. <p>Publish Date: 2018-12-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000872>CVE-2018-1000872</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000872">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000872</a></p> <p>Release Date: 2018-12-20</p> <p>Fix Resolution: v0.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in pykmip tar gz cve medium severity vulnerability vulnerable library pykmip tar gz kmip library library home page a href path to dependency file mongo src third party wiredtiger lang python path to vulnerable library src third party wiredtiger lang python src third party wiredtiger bench workgen dependency hierarchy x pykmip tar gz vulnerable library found in head commit a href found in base branch main vulnerability details openkmip pykmip version all versions before contains a cwe resource management errors similar issue to cve vulnerability in pykmip server that can result in dos the server can be made unavailable by one or more clients opening all of the available sockets this attack appear to be exploitable via a client or clients open sockets with the server and then never close them this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
57,730
16,014,438,988
IssuesEvent
2021-04-20 14:28:47
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
`develop.element.io` with Spaces enabled switching rooms and spaces on its own
T-Defect
### Description After a recent `develop.elements.io` update my `develop.element.io` client with Spaces enabled got stuck in an infinite loop switching between rooms and spaces on its own without any additional user input. I've got a number of private spaces for grouping rooms and the loop involves at least one room included in more than one space and both spaces that include it. The weird switching started before the loop and looked like this: - I click a different space - the space opens and some room opens - space selection goes back to the one I was in - room changes again - result: I'm back in the same room where I were unable to go into rooms in the other space ### Environment `develop.element.io` with Spaces enabled, Firefox Nightly, Linux The problem started after an update that included https://github.com/matrix-org/matrix-react-sdk/pull/5795, which looks like a likely cause, so cc @SimonBrandner @t3chguy
1.0
`develop.element.io` with Spaces enabled switching rooms and spaces on its own - ### Description After a recent `develop.elements.io` update my `develop.element.io` client with Spaces enabled got stuck in an infinite loop switching between rooms and spaces on its own without any additional user input. I've got a number of private spaces for grouping rooms and the loop involves at least one room included in more than one space and both spaces that include it. The weird switching started before the loop and looked like this: - I click a different space - the space opens and some room opens - space selection goes back to the one I was in - room changes again - result: I'm back in the same room where I were unable to go into rooms in the other space ### Environment `develop.element.io` with Spaces enabled, Firefox Nightly, Linux The problem started after an update that included https://github.com/matrix-org/matrix-react-sdk/pull/5795, which looks like a likely cause, so cc @SimonBrandner @t3chguy
defect
develop element io with spaces enabled switching rooms and spaces on its own description after a recent develop elements io update my develop element io client with spaces enabled got stuck in an infinite loop switching between rooms and spaces on its own without any additional user input i ve got a number of private spaces for grouping rooms and the loop involves at least one room included in more than one space and both spaces that include it the weird switching started before the loop and looked like this i click a different space the space opens and some room opens space selection goes back to the one i was in room changes again result i m back in the same room where i were unable to go into rooms in the other space environment develop element io with spaces enabled firefox nightly linux the problem started after an update that included which looks like a likely cause so cc simonbrandner
1
14,475
2,812,852,759
IssuesEvent
2015-05-18 11:26:25
jliljebl/flowblade
https://api.github.com/repos/jliljebl/flowblade
closed
logic of alpha filtering confusing
auto-migrated Priority-Medium Type-Defect
``` When using a compositor (dissolve, ...) the alpha opacity can be set to over/and/or/xor. The result when using 'over' is quit clear. But with the other values, it seems to me that something is going wrong: - when alpha in the source varies from 99-1%, starting from 49% the result is reverted in some way (49% becomes less transparent than 1%) - Using an opacity of 99% may give a significant different result than using 100% opacity - Using an alpha value of 99% in the source may give a significant different result than using 100% These strange results do occur in version 12 and 14. I don't know if it's Flowblade mixing up or if it is MLT. You can find the file I used for testing in the attachment. It's rather difficult to catch the logic behind what is happening. I wanted to create a mask that adds transparency to an underlaying layer. By trial and error I got to the following solution (see also last sequence in attached test project): V3: mask --dissolve(xor,opacity 99)--> V2:layer --dissolve(xor,opacity 100)--> V1:background Anyway, the effects that can be created by mixing opacity are really great. If I only could understand the workings behind ... Bart. ``` Original issue reported on code.google.com by `bart2...@gmail.com` on 5 Aug 2014 at 2:54 Attachments: * [Flowblade alpha test.tar.gz](https://storage.googleapis.com/google-code-attachments/flowblade/issue-92/comment-0/Flowblade alpha test.tar.gz)
1.0
logic of alpha filtering confusing - ``` When using a compositor (dissolve, ...) the alpha opacity can be set to over/and/or/xor. The result when using 'over' is quit clear. But with the other values, it seems to me that something is going wrong: - when alpha in the source varies from 99-1%, starting from 49% the result is reverted in some way (49% becomes less transparent than 1%) - Using an opacity of 99% may give a significant different result than using 100% opacity - Using an alpha value of 99% in the source may give a significant different result than using 100% These strange results do occur in version 12 and 14. I don't know if it's Flowblade mixing up or if it is MLT. You can find the file I used for testing in the attachment. It's rather difficult to catch the logic behind what is happening. I wanted to create a mask that adds transparency to an underlaying layer. By trial and error I got to the following solution (see also last sequence in attached test project): V3: mask --dissolve(xor,opacity 99)--> V2:layer --dissolve(xor,opacity 100)--> V1:background Anyway, the effects that can be created by mixing opacity are really great. If I only could understand the workings behind ... Bart. ``` Original issue reported on code.google.com by `bart2...@gmail.com` on 5 Aug 2014 at 2:54 Attachments: * [Flowblade alpha test.tar.gz](https://storage.googleapis.com/google-code-attachments/flowblade/issue-92/comment-0/Flowblade alpha test.tar.gz)
defect
logic of alpha filtering confusing when using a compositor dissolve the alpha opacity can be set to over and or xor the result when using over is quit clear but with the other values it seems to me that something is going wrong when alpha in the source varies from starting from the result is reverted in some way becomes less transparent than using an opacity of may give a significant different result than using opacity using an alpha value of in the source may give a significant different result than using these strange results do occur in version and i don t know if it s flowblade mixing up or if it is mlt you can find the file i used for testing in the attachment it s rather difficult to catch the logic behind what is happening i wanted to create a mask that adds transparency to an underlaying layer by trial and error i got to the following solution see also last sequence in attached test project mask dissolve xor opacity layer dissolve xor opacity background anyway the effects that can be created by mixing opacity are really great if i only could understand the workings behind bart original issue reported on code google com by gmail com on aug at attachments alpha test tar gz
1
7,940
2,611,067,848
IssuesEvent
2015-02-27 00:31:43
alistairreilly/andors-trail
https://api.github.com/repos/alistairreilly/andors-trail
closed
when dpad-minimizing is disabled - hit the mid to attack
auto-migrated Milestone-0.6.10 Type-Defect
``` Currently it works like this: 1) When dpad-minimizing is disabled: Hit the mid results in minimizing of the dpad. 2) When dpad-minimizing is disabled: Hit the mid results in no action It should work like this: 1) When dpad-minimizing is disabled: Hit the mid results in minimizing of the dpad. 2) When dpad-minimizing is disabled: Hit the mid results in an attack (when in combat mode) Patch attached. ``` Original issue reported on code.google.com by `SamuelPl...@gmail.com` on 5 Oct 2011 at 7:34 Attachments: * [when dpad-minimizing is disabled - hit the mid to attack.patch](https://storage.googleapis.com/google-code-attachments/andors-trail/issue-246/comment-0/when dpad-minimizing is disabled - hit the mid to attack.patch)
1.0
when dpad-minimizing is disabled - hit the mid to attack - ``` Currently it works like this: 1) When dpad-minimizing is disabled: Hit the mid results in minimizing of the dpad. 2) When dpad-minimizing is disabled: Hit the mid results in no action It should work like this: 1) When dpad-minimizing is disabled: Hit the mid results in minimizing of the dpad. 2) When dpad-minimizing is disabled: Hit the mid results in an attack (when in combat mode) Patch attached. ``` Original issue reported on code.google.com by `SamuelPl...@gmail.com` on 5 Oct 2011 at 7:34 Attachments: * [when dpad-minimizing is disabled - hit the mid to attack.patch](https://storage.googleapis.com/google-code-attachments/andors-trail/issue-246/comment-0/when dpad-minimizing is disabled - hit the mid to attack.patch)
defect
when dpad minimizing is disabled hit the mid to attack currently it works like this when dpad minimizing is disabled hit the mid results in minimizing of the dpad when dpad minimizing is disabled hit the mid results in no action it should work like this when dpad minimizing is disabled hit the mid results in minimizing of the dpad when dpad minimizing is disabled hit the mid results in an attack when in combat mode patch attached original issue reported on code google com by samuelpl gmail com on oct at attachments dpad minimizing is disabled hit the mid to attack patch
1
51,361
13,207,442,909
IssuesEvent
2020-08-14 23:07:10
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
trunk of toprec doesn't compile (Trac #283)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/283">https://code.icecube.wisc.edu/projects/icecube/ticket/283</a>, reported by maseand owned by kislat</em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-07T22:32:33", "_ts": "1436308353324715", "description": "Hi,\n\nThe trunk of toprec doesn't compile with following errors.\nCan we changed the max(0.0f, ...) to max(0.0, ...)?\n(With the above change, I could compile the toprec.)\n\n\n[ 78%] Building CXX object toprec/CMakeFiles/toprec.dir/private/toprec/I3TopRecoCurved.cxx.o\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx\n : In member function `virtual void I3TopRecoCurved::Physics(I3FramePtr)':\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx:498:\n error: No match for `max(float, double)'\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx:499:\n error: No match for `max(float, double)'\nmake[2]: *** [toprec/CMakeFiles/toprec.dir/private/toprec/I3TopRecoCurved.cxx.o] Error 1\nmake[1]: *** [toprec/CMakeFiles/toprec.dir/all] Error 2\nmake: *** [all] Error 2\n", "reporter": "mase", "cc": "", "resolution": "invalid", "time": "2011-06-13T06:45:56", "component": "combo reconstruction", "summary": "trunk of toprec doesn't compile", "priority": "major", "keywords": "", "milestone": "", "owner": "kislat", "type": "defect" } ``` </p> </details>
1.0
trunk of toprec doesn't compile (Trac #283) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/283">https://code.icecube.wisc.edu/projects/icecube/ticket/283</a>, reported by maseand owned by kislat</em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-07T22:32:33", "_ts": "1436308353324715", "description": "Hi,\n\nThe trunk of toprec doesn't compile with following errors.\nCan we changed the max(0.0f, ...) to max(0.0, ...)?\n(With the above change, I could compile the toprec.)\n\n\n[ 78%] Building CXX object toprec/CMakeFiles/toprec.dir/private/toprec/I3TopRecoCurved.cxx.o\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx\n : In member function `virtual void I3TopRecoCurved::Physics(I3FramePtr)':\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx:498:\n error: No match for `max(float, double)'\n/home/mase/work/I3/EHE/test/trunk/src/toprec/private/toprec/I3TopRecoCurved.cxx:499:\n error: No match for `max(float, double)'\nmake[2]: *** [toprec/CMakeFiles/toprec.dir/private/toprec/I3TopRecoCurved.cxx.o] Error 1\nmake[1]: *** [toprec/CMakeFiles/toprec.dir/all] Error 2\nmake: *** [all] Error 2\n", "reporter": "mase", "cc": "", "resolution": "invalid", "time": "2011-06-13T06:45:56", "component": "combo reconstruction", "summary": "trunk of toprec doesn't compile", "priority": "major", "keywords": "", "milestone": "", "owner": "kislat", "type": "defect" } ``` </p> </details>
defect
trunk of toprec doesn t compile trac migrated from json status closed changetime ts description hi n nthe trunk of toprec doesn t compile with following errors ncan we changed the max to max n with the above change i could compile the toprec n n n building cxx object toprec cmakefiles toprec dir private toprec cxx o n home mase work ehe test trunk src toprec private toprec cxx n in member function virtual void physics n home mase work ehe test trunk src toprec private toprec cxx n error no match for max float double n home mase work ehe test trunk src toprec private toprec cxx n error no match for max float double nmake error nmake error nmake error n reporter mase cc resolution invalid time component combo reconstruction summary trunk of toprec doesn t compile priority major keywords milestone owner kislat type defect
1
221,900
7,398,722,560
IssuesEvent
2018-03-19 07:47:59
kubernetes/federation
https://api.github.com/repos/kubernetes/federation
closed
Federated deployments cannot be deleted
kind/bug priority/important-soon sig/cli sig/multicluster
<a href="https://github.com/jsgoller1"><img src="https://avatars3.githubusercontent.com/u/1567977?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [jsgoller1](https://github.com/jsgoller1)** _Sunday Oct 08, 2017 at 13:20 GMT_ _Originally opened as https://github.com/kubernetes/kubernetes/issues/53566_ ---- **Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug /sig cli /sig federation @kubernetes/sig-federation **What happened**: I have a federated cluster composed of one host cluster and one joined cluster running in AWS. They were created following the steps in the documentation, and I am able to see via the kubernetes dashboard that all components are running correctly. To test the federated setup, I used a simple nginx deployment also from the docs, stored in `deployment.proxy.yml`: ``` apiVersion: extensions/v1beta1 kind: Deployment metadata: name: proxy spec: replicas: 5 # tells deployment to run 2 pods matching the template template: # create pods using pod definition in this template metadata: labels: app: nginx spec: containers: - name: nginx image: nginx:latest ports: - containerPort: 80 ``` I then execute `kubectl create --context=federation -f deployment.proxy.yml`. Assuming I don't change `apiVersion`, the deployment runs fine and almost immediately reaches the desired count (if I use `apiVersion: apps/v1beta2` or any other API version than `extensions/` with my federated setup, I get errors and it won't start). I then go to tear down the deployment by executing `kubectl delete deployment --context=federation proxy`. However, kubectl reports `error: timed out waiting for the condition`. The deployment instances spin down to zero, but the deployment won't disappear: ``` NAME DESIRED CURRENT UP-TO-DATE AVAILABLE AGE proxy 0 0 0 0 1h ``` I can delete them via the dashboard, or by switching to joined cluster's context and using `kubctl delete deployment proxy`, but even after kubectll reports they are deleted successfully, they reappear and still cannot be deleted from the federation context. **What you expected to happen**: The deployment to be deleted. This works fine if I'm not using my federated context. **How to reproduce it (as minimally and precisely as possible)**: Create a federated setup via the docs, launch the above job, and then try to delete it. **Environment**: - Kubernetes version: ``` Client Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.0", GitCommit:"6e937839ac04a38cac63e6a7a306c5d035fe7b0a", GitTreeState:"clean", BuildDate:"2017-09-28T22:57:57Z", GoVersion:"go1.8.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", ``` - Cloud provider or hardware configuration: AWS - OS (e.g. from /etc/os-release): Ubuntu 16.04
1.0
Federated deployments cannot be deleted - <a href="https://github.com/jsgoller1"><img src="https://avatars3.githubusercontent.com/u/1567977?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [jsgoller1](https://github.com/jsgoller1)** _Sunday Oct 08, 2017 at 13:20 GMT_ _Originally opened as https://github.com/kubernetes/kubernetes/issues/53566_ ---- **Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug /sig cli /sig federation @kubernetes/sig-federation **What happened**: I have a federated cluster composed of one host cluster and one joined cluster running in AWS. They were created following the steps in the documentation, and I am able to see via the kubernetes dashboard that all components are running correctly. To test the federated setup, I used a simple nginx deployment also from the docs, stored in `deployment.proxy.yml`: ``` apiVersion: extensions/v1beta1 kind: Deployment metadata: name: proxy spec: replicas: 5 # tells deployment to run 2 pods matching the template template: # create pods using pod definition in this template metadata: labels: app: nginx spec: containers: - name: nginx image: nginx:latest ports: - containerPort: 80 ``` I then execute `kubectl create --context=federation -f deployment.proxy.yml`. Assuming I don't change `apiVersion`, the deployment runs fine and almost immediately reaches the desired count (if I use `apiVersion: apps/v1beta2` or any other API version than `extensions/` with my federated setup, I get errors and it won't start). I then go to tear down the deployment by executing `kubectl delete deployment --context=federation proxy`. However, kubectl reports `error: timed out waiting for the condition`. The deployment instances spin down to zero, but the deployment won't disappear: ``` NAME DESIRED CURRENT UP-TO-DATE AVAILABLE AGE proxy 0 0 0 0 1h ``` I can delete them via the dashboard, or by switching to joined cluster's context and using `kubctl delete deployment proxy`, but even after kubectll reports they are deleted successfully, they reappear and still cannot be deleted from the federation context. **What you expected to happen**: The deployment to be deleted. This works fine if I'm not using my federated context. **How to reproduce it (as minimally and precisely as possible)**: Create a federated setup via the docs, launch the above job, and then try to delete it. **Environment**: - Kubernetes version: ``` Client Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.0", GitCommit:"6e937839ac04a38cac63e6a7a306c5d035fe7b0a", GitTreeState:"clean", BuildDate:"2017-09-28T22:57:57Z", GoVersion:"go1.8.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", ``` - Cloud provider or hardware configuration: AWS - OS (e.g. from /etc/os-release): Ubuntu 16.04
non_defect
federated deployments cannot be deleted issue by sunday oct at gmt originally opened as is this a bug report or feature request kind bug sig cli sig federation kubernetes sig federation what happened i have a federated cluster composed of one host cluster and one joined cluster running in aws they were created following the steps in the documentation and i am able to see via the kubernetes dashboard that all components are running correctly to test the federated setup i used a simple nginx deployment also from the docs stored in deployment proxy yml apiversion extensions kind deployment metadata name proxy spec replicas tells deployment to run pods matching the template template create pods using pod definition in this template metadata labels app nginx spec containers name nginx image nginx latest ports containerport i then execute kubectl create context federation f deployment proxy yml assuming i don t change apiversion the deployment runs fine and almost immediately reaches the desired count if i use apiversion apps or any other api version than extensions with my federated setup i get errors and it won t start i then go to tear down the deployment by executing kubectl delete deployment context federation proxy however kubectl reports error timed out waiting for the condition the deployment instances spin down to zero but the deployment won t disappear name desired current up to date available age proxy i can delete them via the dashboard or by switching to joined cluster s context and using kubctl delete deployment proxy but even after kubectll reports they are deleted successfully they reappear and still cannot be deleted from the federation context what you expected to happen the deployment to be deleted this works fine if i m not using my federated context how to reproduce it as minimally and precisely as possible create a federated setup via the docs launch the above job and then try to delete it environment kubernetes version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major cloud provider or hardware configuration aws os e g from etc os release ubuntu
0
15,475
2,856,413,914
IssuesEvent
2015-06-02 14:52:37
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
JS error in Dialog
5.1.19 5.2.6 defect
A Javascript bug has been introduced when searching for existing dialog by id, the value is not escaped (":"), bringing up: Uncaught Error: Syntax error, unrecognized expression: [id=form:j_idt397:j_idt400]
1.0
JS error in Dialog - A Javascript bug has been introduced when searching for existing dialog by id, the value is not escaped (":"), bringing up: Uncaught Error: Syntax error, unrecognized expression: [id=form:j_idt397:j_idt400]
defect
js error in dialog a javascript bug has been introduced when searching for existing dialog by id the value is not escaped bringing up uncaught error syntax error unrecognized expression
1
316,553
23,638,011,613
IssuesEvent
2022-08-25 14:42:38
Big-Life-Lab/PHES-ODM
https://api.github.com/repos/Big-Life-Lab/PHES-ODM
opened
Review approach to record time-specific measure
documentation for discussion
There are different options for recording time-specific measures such as flow rates. Provide documentation and examples of these options and review to ensure all options can be recorded.
1.0
Review approach to record time-specific measure - There are different options for recording time-specific measures such as flow rates. Provide documentation and examples of these options and review to ensure all options can be recorded.
non_defect
review approach to record time specific measure there are different options for recording time specific measures such as flow rates provide documentation and examples of these options and review to ensure all options can be recorded
0
73,829
24,815,422,114
IssuesEvent
2022-10-25 12:49:16
bellingard/sonar-scanner-npm
https://api.github.com/repos/bellingard/sonar-scanner-npm
closed
2.8.2 can't download via proxy
help wanted confirmed defect
With version 2.8.1, the scanner download code allows you to set the `https_proxy` / `http_proxy` env vars to specify a proxy to use to download. Version 2.8.2 does not respect those env vars for proxy downloads. I assume this is probably due to the switch from the `download` module in 2.8.1 to `node-downloader-helper` in 2.8.2. Example script that works with 2.8.1 but fails with 2.8.2: ```sh export https_proxy=<myproxy> export http_proxy=<myproxy> npm install -g sonarqube-scanner sonar-scanner -v ``` Can we please get support for proxies restored in the downloader?
1.0
2.8.2 can't download via proxy - With version 2.8.1, the scanner download code allows you to set the `https_proxy` / `http_proxy` env vars to specify a proxy to use to download. Version 2.8.2 does not respect those env vars for proxy downloads. I assume this is probably due to the switch from the `download` module in 2.8.1 to `node-downloader-helper` in 2.8.2. Example script that works with 2.8.1 but fails with 2.8.2: ```sh export https_proxy=<myproxy> export http_proxy=<myproxy> npm install -g sonarqube-scanner sonar-scanner -v ``` Can we please get support for proxies restored in the downloader?
defect
can t download via proxy with version the scanner download code allows you to set the https proxy http proxy env vars to specify a proxy to use to download version does not respect those env vars for proxy downloads i assume this is probably due to the switch from the download module in to node downloader helper in example script that works with but fails with sh export https proxy export http proxy npm install g sonarqube scanner sonar scanner v can we please get support for proxies restored in the downloader
1
15,404
9,997,433,625
IssuesEvent
2019-07-12 04:25:25
flutter/flutter-intellij
https://api.github.com/repos/flutter/flutter-intellij
closed
timing issues w/ initial attempt to list samples
investigate topic-project-creation usability
Follow-up from https://github.com/flutter/flutter-intellij/issues/3404#issuecomment-486848338, where @stevemessick notes: > This happened while connected to the corp network. I see it at home, but that's not surprising. My home network is slow. Corp should be fast. > <img width="308" alt="Screen Shot 2019-04-25 at 2 38 26 PM" src="https://user-images.githubusercontent.com/8518285/56770083-f028aa00-6767-11e9-8cb9-70187c0e7247.png"> The initial listing can take a while (if, for example, the flutter tool triggers SDK and engine downloads). This can be reflected in the UI as an empty list of sample selections.
True
timing issues w/ initial attempt to list samples - Follow-up from https://github.com/flutter/flutter-intellij/issues/3404#issuecomment-486848338, where @stevemessick notes: > This happened while connected to the corp network. I see it at home, but that's not surprising. My home network is slow. Corp should be fast. > <img width="308" alt="Screen Shot 2019-04-25 at 2 38 26 PM" src="https://user-images.githubusercontent.com/8518285/56770083-f028aa00-6767-11e9-8cb9-70187c0e7247.png"> The initial listing can take a while (if, for example, the flutter tool triggers SDK and engine downloads). This can be reflected in the UI as an empty list of sample selections.
non_defect
timing issues w initial attempt to list samples follow up from where stevemessick notes this happened while connected to the corp network i see it at home but that s not surprising my home network is slow corp should be fast img width alt screen shot at pm src the initial listing can take a while if for example the flutter tool triggers sdk and engine downloads this can be reflected in the ui as an empty list of sample selections
0
71,570
23,697,995,345
IssuesEvent
2022-08-29 16:13:54
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
closed
App shows blank screen only with element icon after login
T-Defect A-Startup S-Critical O-Occasional
### Steps to reproduce 1. Where are you starting? What can you see? Opening the app, I am seeing only a blank screen with the element logo ![signal-2021-12-08-16-35-02-856](https://user-images.githubusercontent.com/88454888/145237705-985176f6-3057-407f-96d5-c5ee458bdbad.png) 2. What do you click? I can click nothing ### Outcome #### What did you expect? Seeing pin prompt after opening the app #### What happened instead? Blank screen appeared, nothing was possible but reinstalling the app. Did not fix the issue after login in again. ### Your phone model iPhone XR ### Operating system version iOS 15.1 ### Application version 1.6.8 ### Homeserver --> pm ### Will you send logs? Yes
1.0
App shows blank screen only with element icon after login - ### Steps to reproduce 1. Where are you starting? What can you see? Opening the app, I am seeing only a blank screen with the element logo ![signal-2021-12-08-16-35-02-856](https://user-images.githubusercontent.com/88454888/145237705-985176f6-3057-407f-96d5-c5ee458bdbad.png) 2. What do you click? I can click nothing ### Outcome #### What did you expect? Seeing pin prompt after opening the app #### What happened instead? Blank screen appeared, nothing was possible but reinstalling the app. Did not fix the issue after login in again. ### Your phone model iPhone XR ### Operating system version iOS 15.1 ### Application version 1.6.8 ### Homeserver --> pm ### Will you send logs? Yes
defect
app shows blank screen only with element icon after login steps to reproduce where are you starting what can you see opening the app i am seeing only a blank screen with the element logo what do you click i can click nothing outcome what did you expect seeing pin prompt after opening the app what happened instead blank screen appeared nothing was possible but reinstalling the app did not fix the issue after login in again your phone model iphone xr operating system version ios application version homeserver pm will you send logs yes
1
170,632
14,266,261,749
IssuesEvent
2020-11-20 18:24:28
MLH-Fellowship/OSS-Guide
https://api.github.com/repos/MLH-Fellowship/OSS-Guide
closed
Planning
documentation enhancement front-end
# Update the General Guide ## Behavior Create an MDX doc which explains the section in depth with appropriate examples and add it in the appropriate section in the docs page. - Planning
1.0
Planning - # Update the General Guide ## Behavior Create an MDX doc which explains the section in depth with appropriate examples and add it in the appropriate section in the docs page. - Planning
non_defect
planning update the general guide behavior create an mdx doc which explains the section in depth with appropriate examples and add it in the appropriate section in the docs page planning
0
7,177
2,610,355,832
IssuesEvent
2015-02-26 19:55:14
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Post is published as page instead.
auto-migrated Priority-Medium Type-Defect
``` What's the problem? Creating a publishing a post to my Wordpress.com blog works fine except that the post is published as a page instead. What browser are you using? I have encountered the problem in both Firefox and Chrome browsers. What version of ScribeFire are you running? 1.9 ``` ----- Original issue reported on code.google.com by `linus.fe...@gmail.com` on 9 Oct 2011 at 3:30
1.0
Post is published as page instead. - ``` What's the problem? Creating a publishing a post to my Wordpress.com blog works fine except that the post is published as a page instead. What browser are you using? I have encountered the problem in both Firefox and Chrome browsers. What version of ScribeFire are you running? 1.9 ``` ----- Original issue reported on code.google.com by `linus.fe...@gmail.com` on 9 Oct 2011 at 3:30
defect
post is published as page instead what s the problem creating a publishing a post to my wordpress com blog works fine except that the post is published as a page instead what browser are you using i have encountered the problem in both firefox and chrome browsers what version of scribefire are you running original issue reported on code google com by linus fe gmail com on oct at
1
17,847
10,133,704,608
IssuesEvent
2019-08-02 04:38:19
okTurtles/group-income-simple
https://api.github.com/repos/okTurtles/group-income-simple
opened
Handle DoS from malformed messages sent by group members
App:Frontend Kind:Core Level:Advanced Note:Security
### Problem If a client of a group member sends a poorly formed action to the GroupContract, it will be rejected by all other members, however, the server will update the contract HEAD to point to this message, which should've been discarded (but the server has no way of knowing that). So all clients will become stuck at this HEAD and unable to move forward. ### Solution The other (functioning, non-broken) clients should: - Ignore the malformed message, not allowing it to modify the internal state at all - Identify the user responsible and automatically initiate a proposal to remove them from the group ("auto-banning") - Continue building the chain on top of that message as if nothing happened
True
Handle DoS from malformed messages sent by group members - ### Problem If a client of a group member sends a poorly formed action to the GroupContract, it will be rejected by all other members, however, the server will update the contract HEAD to point to this message, which should've been discarded (but the server has no way of knowing that). So all clients will become stuck at this HEAD and unable to move forward. ### Solution The other (functioning, non-broken) clients should: - Ignore the malformed message, not allowing it to modify the internal state at all - Identify the user responsible and automatically initiate a proposal to remove them from the group ("auto-banning") - Continue building the chain on top of that message as if nothing happened
non_defect
handle dos from malformed messages sent by group members problem if a client of a group member sends a poorly formed action to the groupcontract it will be rejected by all other members however the server will update the contract head to point to this message which should ve been discarded but the server has no way of knowing that so all clients will become stuck at this head and unable to move forward solution the other functioning non broken clients should ignore the malformed message not allowing it to modify the internal state at all identify the user responsible and automatically initiate a proposal to remove them from the group auto banning continue building the chain on top of that message as if nothing happened
0
50,890
12,601,849,697
IssuesEvent
2020-06-11 10:35:51
canonical-web-and-design/build.snapcraft.io
https://api.github.com/repos/canonical-web-and-design/build.snapcraft.io
closed
“To test this build” text is inaccurate
Build page Design: Required Test instructions
[Split off from #641] Currently on individual build pages, for published builds, we invite you to install that build for testing: ![“To test this build on your PC or cloud instance: … The installed snap will not be updated.”](https://cloud.githubusercontent.com/assets/1523179/25176665/de20109c-24f6-11e7-90a7-bcd35065f4bd.png) Unfortunately, this is both inaccurate and incomplete: <table> <tr><th>What it claims/implies</th><th>What currently happens</th><th>What snapd developers are proposing</th></tr> <tr><td>Publishers/collaborators can install any specific revision ever published</td><td>? <strong>(A)</strong></td><td>Publishers/collaborators can install any specific revision ever published, except that they can install a revision before their currently-installed one <em>only if they had previously installed it</em></td></tr> <tr><td>Non-publishers + non-collaborators can also install any specific revision ever published</td><td>? <strong>(D)</strong></td><td>Non-publishers/collaborators can’t install past revisions at all</td></tr> <tr><td>If you install a specific revision, it will not be updated automatically</td><td>If you install a specific revision, it will be updated within ~6 hours of a new revision being released</td><td>If you install a specific revision, it will be updated within ~6 hours of a new revision being released</td></tr> <tr><td>—</td><td>Short of disconnecting from the Internet, there’s no way to stop the revision from being updated <strong>(J)</strong></td><td>If you want to delay your specific revision being replaced by a new revision, you should use a “postpone any refreshes by N hours” feature</td></tr> <tr><td>—</td><td>To resume tracking latest without a new revision being released, uninstall+reinstall the snap <strong>(K)</strong></td><td>To prepone the postponed refresh, … ?</td></tr> </table> While we wait for the proposals to be implemented, we should fix the text to match current reality. This requires answering these questions: (A) Which revisions can publishers/collaborators currently install? Is there any limit on installing a revision if you have a newer revision currently installed? (D) Which revisions can non-publishers + non-collaborators currently install? Is there any limit on installing a revision if you have a newer revision currently installed? (J) Is this correct? Is there any kinder way of blocking updates to a particular snap? (K) Is this correct? Is there any simpler way of updating to latest?
1.0
“To test this build” text is inaccurate - [Split off from #641] Currently on individual build pages, for published builds, we invite you to install that build for testing: ![“To test this build on your PC or cloud instance: … The installed snap will not be updated.”](https://cloud.githubusercontent.com/assets/1523179/25176665/de20109c-24f6-11e7-90a7-bcd35065f4bd.png) Unfortunately, this is both inaccurate and incomplete: <table> <tr><th>What it claims/implies</th><th>What currently happens</th><th>What snapd developers are proposing</th></tr> <tr><td>Publishers/collaborators can install any specific revision ever published</td><td>? <strong>(A)</strong></td><td>Publishers/collaborators can install any specific revision ever published, except that they can install a revision before their currently-installed one <em>only if they had previously installed it</em></td></tr> <tr><td>Non-publishers + non-collaborators can also install any specific revision ever published</td><td>? <strong>(D)</strong></td><td>Non-publishers/collaborators can’t install past revisions at all</td></tr> <tr><td>If you install a specific revision, it will not be updated automatically</td><td>If you install a specific revision, it will be updated within ~6 hours of a new revision being released</td><td>If you install a specific revision, it will be updated within ~6 hours of a new revision being released</td></tr> <tr><td>—</td><td>Short of disconnecting from the Internet, there’s no way to stop the revision from being updated <strong>(J)</strong></td><td>If you want to delay your specific revision being replaced by a new revision, you should use a “postpone any refreshes by N hours” feature</td></tr> <tr><td>—</td><td>To resume tracking latest without a new revision being released, uninstall+reinstall the snap <strong>(K)</strong></td><td>To prepone the postponed refresh, … ?</td></tr> </table> While we wait for the proposals to be implemented, we should fix the text to match current reality. This requires answering these questions: (A) Which revisions can publishers/collaborators currently install? Is there any limit on installing a revision if you have a newer revision currently installed? (D) Which revisions can non-publishers + non-collaborators currently install? Is there any limit on installing a revision if you have a newer revision currently installed? (J) Is this correct? Is there any kinder way of blocking updates to a particular snap? (K) Is this correct? Is there any simpler way of updating to latest?
non_defect
“to test this build” text is inaccurate currently on individual build pages for published builds we invite you to install that build for testing unfortunately this is both inaccurate and incomplete what it claims implies what currently happens what snapd developers are proposing publishers collaborators can install any specific revision ever published a publishers collaborators can install any specific revision ever published except that they can install a revision before their currently installed one only if they had previously installed it non publishers non collaborators can also install any specific revision ever published d non publishers collaborators can’t install past revisions at all if you install a specific revision it will not be updated automatically if you install a specific revision it will be updated within hours of a new revision being released if you install a specific revision it will be updated within hours of a new revision being released — short of disconnecting from the internet there’s no way to stop the revision from being updated j if you want to delay your specific revision being replaced by a new revision you should use a “postpone any refreshes by n hours” feature — to resume tracking latest without a new revision being released uninstall reinstall the snap k to prepone the postponed refresh … while we wait for the proposals to be implemented we should fix the text to match current reality this requires answering these questions a which revisions can publishers collaborators currently install is there any limit on installing a revision if you have a newer revision currently installed d which revisions can non publishers non collaborators currently install is there any limit on installing a revision if you have a newer revision currently installed j is this correct is there any kinder way of blocking updates to a particular snap k is this correct is there any simpler way of updating to latest
0
161,457
20,154,029,268
IssuesEvent
2022-02-09 14:58:03
kapseliboi/bui
https://api.github.com/repos/kapseliboi/bui
opened
CVE-2022-21670 (Medium) detected in markdown-it-8.4.2.tgz
security vulnerability
## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-8.4.2.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - docpress-0.7.4.tgz (Root Library) - docpress-core-0.9.0.tgz - :x: **markdown-it-8.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/bui/commit/e8304e6335e5d45f5599a6dd9950348f734192b7">e8304e6335e5d45f5599a6dd9950348f734192b7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: markdown-it - 12.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21670 (Medium) detected in markdown-it-8.4.2.tgz - ## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-8.4.2.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-8.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - docpress-0.7.4.tgz (Root Library) - docpress-core-0.9.0.tgz - :x: **markdown-it-8.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/bui/commit/e8304e6335e5d45f5599a6dd9950348f734192b7">e8304e6335e5d45f5599a6dd9950348f734192b7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: markdown-it - 12.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in markdown it tgz cve medium severity vulnerability vulnerable library markdown it tgz markdown it modern pluggable markdown parser library home page a href path to dependency file package json path to vulnerable library node modules markdown it package json dependency hierarchy docpress tgz root library docpress core tgz x markdown it tgz vulnerable library found in head commit a href found in base branch master vulnerability details markdown it is a markdown parser prior to version special patterns with length greater than thousand characterss could slow down the parser significantly users should upgrade to version to receive a patch there are no known workarounds aside from upgrading publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution markdown it step up your open source security game with whitesource
0
70,375
23,145,167,974
IssuesEvent
2022-07-28 23:22:59
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
fails to build on kernel 5.16, rocky linux, standard kernel-ml package
Type: Defect
CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx2.o CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512f.o CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512bw.o /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c: In function 'zvol_os_create_minor': /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c:1075:3: error: ignoring return value of 'add_disk', declared with attribute warn_unused_result [-Werror=unused-result] add_disk(zv->zv_zso->zvo_disk); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors make[5]: *** [scripts/Makefile.build:287: /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.o] Error 1 make[5]: *** Waiting for unfinished jobs.... LD [M] /var/lib/dkms/zfs/2.0.7/build/module/zstd/zzstd.o make[4]: *** [scripts/Makefile.build:549: /var/lib/dkms/zfs/2.0.7/build/module/zfs] Error 2 make[3]: *** [Makefile:1846: /var/lib/dkms/zfs/2.0.7/build/module] Error 2 make[3]: Leaving directory '/usr/src/kernels/5.16.5-1.el8.elrepo.x86_64' make[2]: *** [Makefile:49: modules-Linux] Error 2 make[2]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build/module' make[1]: *** [Makefile:898: all-recursive] Error 1 make[1]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build' make: *** [Makefile:759: all] Error 2 router:/usr/src/zfs-2.0.7 # uname -a Linux router 5.16.5-1.el8.elrepo.x86_64 #1 SMP PREEMPT Mon Jan 31 19:03:05 EST 2022 x86_64 x86_64 x86_64 GNU/Linux router:/usr/src/zfs-2.0.7 # cat /etc/redhat-release Rocky Linux release 8.5 (Green Obsidian) router:/usr/src/zfs-2.0.7 #
1.0
fails to build on kernel 5.16, rocky linux, standard kernel-ml package - CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx2.o CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512f.o CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512bw.o /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c: In function 'zvol_os_create_minor': /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c:1075:3: error: ignoring return value of 'add_disk', declared with attribute warn_unused_result [-Werror=unused-result] add_disk(zv->zv_zso->zvo_disk); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors make[5]: *** [scripts/Makefile.build:287: /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.o] Error 1 make[5]: *** Waiting for unfinished jobs.... LD [M] /var/lib/dkms/zfs/2.0.7/build/module/zstd/zzstd.o make[4]: *** [scripts/Makefile.build:549: /var/lib/dkms/zfs/2.0.7/build/module/zfs] Error 2 make[3]: *** [Makefile:1846: /var/lib/dkms/zfs/2.0.7/build/module] Error 2 make[3]: Leaving directory '/usr/src/kernels/5.16.5-1.el8.elrepo.x86_64' make[2]: *** [Makefile:49: modules-Linux] Error 2 make[2]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build/module' make[1]: *** [Makefile:898: all-recursive] Error 1 make[1]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build' make: *** [Makefile:759: all] Error 2 router:/usr/src/zfs-2.0.7 # uname -a Linux router 5.16.5-1.el8.elrepo.x86_64 #1 SMP PREEMPT Mon Jan 31 19:03:05 EST 2022 x86_64 x86_64 x86_64 GNU/Linux router:/usr/src/zfs-2.0.7 # cat /etc/redhat-release Rocky Linux release 8.5 (Green Obsidian) router:/usr/src/zfs-2.0.7 #
defect
fails to build on kernel rocky linux standard kernel ml package cc var lib dkms zfs build module zfs vdev raidz math o cc var lib dkms zfs build module zfs vdev raidz math o cc var lib dkms zfs build module zfs vdev raidz math o var lib dkms zfs build module zfs os linux zfs zvol os c in function zvol os create minor var lib dkms zfs build module zfs os linux zfs zvol os c error ignoring return value of add disk declared with attribute warn unused result add disk zv zv zso zvo disk all warnings being treated as errors make error make waiting for unfinished jobs ld var lib dkms zfs build module zstd zzstd o make error make error make leaving directory usr src kernels elrepo make error make leaving directory var lib dkms zfs build module make error make leaving directory var lib dkms zfs build make error router usr src zfs uname a linux router elrepo smp preempt mon jan est gnu linux router usr src zfs cat etc redhat release rocky linux release green obsidian router usr src zfs
1
819,388
30,733,247,064
IssuesEvent
2023-07-28 04:55:41
Bentheburrito/cai
https://api.github.com/repos/Bentheburrito/cai
closed
Event Feed: Character loadout icons
feature good first issue low priority
Some events contain `loadout_id` fields. Creating a map under `CAI` of the form `%{loadout_id => image_url}` would let the event feed render a loadout/class icon next to character names.
1.0
Event Feed: Character loadout icons - Some events contain `loadout_id` fields. Creating a map under `CAI` of the form `%{loadout_id => image_url}` would let the event feed render a loadout/class icon next to character names.
non_defect
event feed character loadout icons some events contain loadout id fields creating a map under cai of the form loadout id image url would let the event feed render a loadout class icon next to character names
0
149,381
19,577,949,646
IssuesEvent
2022-01-04 17:21:32
ibm-cio-vulnerability-scanning/insomnia
https://api.github.com/repos/ibm-cio-vulnerability-scanning/insomnia
closed
CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz - autoclosed
security vulnerability
## CVE-2021-27515 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p> Dependency Hierarchy: - swagger-ui-react-3.30.2.tgz (Root Library) - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-cio-vulnerability-scanning/insomnia/commit/66be0bf823b886e28f10e13043b244516268423b">66be0bf823b886e28f10e13043b244516268423b</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path. <p>Publish Date: 2021-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p> <p>Release Date: 2021-02-22</p> <p>Fix Resolution (url-parse): 1.5.0</p> <p>Direct dependency fix Resolution (swagger-ui-react): 3.31.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz - autoclosed - ## CVE-2021-27515 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p> Dependency Hierarchy: - swagger-ui-react-3.30.2.tgz (Root Library) - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-cio-vulnerability-scanning/insomnia/commit/66be0bf823b886e28f10e13043b244516268423b">66be0bf823b886e28f10e13043b244516268423b</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path. <p>Publish Date: 2021-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p> <p>Release Date: 2021-02-22</p> <p>Fix Resolution (url-parse): 1.5.0</p> <p>Direct dependency fix Resolution (swagger-ui-react): 3.31.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in url parse tgz autoclosed cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href dependency hierarchy swagger ui react tgz root library x url parse tgz vulnerable library found in head commit a href found in base branch develop vulnerability details url parse before mishandles certain uses of backslash such as http and interprets the uri as a relative path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution swagger ui react step up your open source security game with whitesource
0
62,447
17,023,924,810
IssuesEvent
2021-07-03 04:35:17
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Specific city boundary relations not showing up in Nominatim
Component: nominatim Priority: major Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 2.45pm, Monday, 24th August 2015]** I have been creating admin_level 8 boundaries for a region and while some have been recognized by Nominatim, others haven't. I. e., some show up in search, others don't; I can only reach them navigating through relations. It is important to mention that in any of the non-recognized regions did I ignore any warnings in JOSM (besides unnamed nodes), and that none of them appear to be faulty to the osm inspector http://tools.geofabrik.de/osmi/ The regions in question are given their relation ids: Ro Segundo: 5458713 San Antonio: 5459859 Turrcares: 5459912 La Garita: 5459924
1.0
Specific city boundary relations not showing up in Nominatim - **[Submitted to the original trac issue database at 2.45pm, Monday, 24th August 2015]** I have been creating admin_level 8 boundaries for a region and while some have been recognized by Nominatim, others haven't. I. e., some show up in search, others don't; I can only reach them navigating through relations. It is important to mention that in any of the non-recognized regions did I ignore any warnings in JOSM (besides unnamed nodes), and that none of them appear to be faulty to the osm inspector http://tools.geofabrik.de/osmi/ The regions in question are given their relation ids: Ro Segundo: 5458713 San Antonio: 5459859 Turrcares: 5459912 La Garita: 5459924
defect
specific city boundary relations not showing up in nominatim i have been creating admin level boundaries for a region and while some have been recognized by nominatim others haven t i e some show up in search others don t i can only reach them navigating through relations it is important to mention that in any of the non recognized regions did i ignore any warnings in josm besides unnamed nodes and that none of them appear to be faulty to the osm inspector the regions in question are given their relation ids ro segundo san antonio turrcares la garita
1
75,100
7,460,009,495
IssuesEvent
2018-03-30 17:47:03
prasadtalasila/IRCLogParser
https://api.github.com/repos/prasadtalasila/IRCLogParser
closed
functional test for Channel Profile
functional test
Create functional tests for verifying the channel profile feature. The relevant lines of code are: **Channel profile (channelprofile.py)** * activity_graph ---> tabular form of figs 3(a) and 3(b) * 27,28,34-38 of [slack.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/slack.py#L34-L38) * 121-125 of [ubuntu.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/ubuntu.py#L121-L125) * 24,25,31-35 of [scummvm.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/scummvm.py#L31-L35) * conversation characteristics ---> Table-II, pg 9 * 24,25,47-64 of [scummvm.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/scummvm.py#L47-L64) * 16-43 of [ubuntu.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/ubuntu.py#L16-L43)
1.0
functional test for Channel Profile - Create functional tests for verifying the channel profile feature. The relevant lines of code are: **Channel profile (channelprofile.py)** * activity_graph ---> tabular form of figs 3(a) and 3(b) * 27,28,34-38 of [slack.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/slack.py#L34-L38) * 121-125 of [ubuntu.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/ubuntu.py#L121-L125) * 24,25,31-35 of [scummvm.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/scummvm.py#L31-L35) * conversation characteristics ---> Table-II, pg 9 * 24,25,47-64 of [scummvm.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/scummvm.py#L47-L64) * 16-43 of [ubuntu.py](https://github.com/prasadtalasila/IRCLogParser/blob/master/ubuntu.py#L16-L43)
non_defect
functional test for channel profile create functional tests for verifying the channel profile feature the relevant lines of code are channel profile channelprofile py activity graph tabular form of figs a and b of of of conversation characteristics table ii pg of of
0
23,577
3,851,864,171
IssuesEvent
2016-04-06 05:27:14
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
latency problem mame4droid 0.139u1
auto-migrated Priority-Medium Type-Defect
``` There's a huge lag choppy problem caused by sound on Mame4Droid 0.139u1 when I played metal slug on the old name4droid 1.5.2 it played fluently and on the new 0.139u1 it freezes and the sound is choppy regardless of the options the game freezes and crashes some times. my ram is 512 mb. cpu 1ghz can I get some help on this plz and great app by the way love it What version of the product are you using? On what operating system? mame4droid 0,139u1. Android OS 4.1.2 jellybean ``` Original issue reported on code.google.com by `Evolutio...@gmail.com` on 18 Jan 2014 at 8:36
1.0
latency problem mame4droid 0.139u1 - ``` There's a huge lag choppy problem caused by sound on Mame4Droid 0.139u1 when I played metal slug on the old name4droid 1.5.2 it played fluently and on the new 0.139u1 it freezes and the sound is choppy regardless of the options the game freezes and crashes some times. my ram is 512 mb. cpu 1ghz can I get some help on this plz and great app by the way love it What version of the product are you using? On what operating system? mame4droid 0,139u1. Android OS 4.1.2 jellybean ``` Original issue reported on code.google.com by `Evolutio...@gmail.com` on 18 Jan 2014 at 8:36
defect
latency problem there s a huge lag choppy problem caused by sound on when i played metal slug on the old it played fluently and on the new it freezes and the sound is choppy regardless of the options the game freezes and crashes some times my ram is mb cpu can i get some help on this plz and great app by the way love it what version of the product are you using on what operating system android os jellybean original issue reported on code google com by evolutio gmail com on jan at
1
192,954
15,363,264,970
IssuesEvent
2021-03-01 20:34:06
AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet
https://api.github.com/repos/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet
closed
AudienceValidator evaluation during TokenValidation doesn't honor ValidateAudience flag
Customer reported Documentation
We're upgrading from an older version 5.6.0 to 6.8.0. Seems like in the newest version, there was a change made to have AudienceValidator throw an exception here: https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/pull/36/commits/70c23ef985ad7d5c20a8e609b6c329642f311786 Which is fine and all. However, now it seems like if you specify a value for AudienceValidator AND set ValidateAudience = false, ValidateAudience does not get honored and we still evaluate the audience validator delegate rather than honoring the ValidateAudience flag. Seems like ValidateAudience needs to be accounted for. If it's set to false, we shouldn't throw on the delegate? We've worked around it for now by just manually nulling out AudienceValidator but still seems like a bug that should be fixed. Sample code: TokenValidationParameters tokenValidationParams = new TokenValidationParameters { AudienceValidator = AudienceValidator, }; tokenValidationParams.ValidateAudience = false; JwtSecurityTokenHandler.ValidateToken( token, tokenValidationParams, out validatedToken); Previous behavior: Passes validation New behavior: Throws Exception: Microsoft.Online.Common.Authentication.SecurityTokenValidation.Exceptions.MalformedTokenException: Malformed access token ---> Microsoft.IdentityModel.Tokens.SecurityTokenInvalidAudienceException: IDX10231: Audience validation failed. Delegate returned false, securitytoken: 'System.IdentityModel.Tokens.Jwt.JwtSecurityToken'.
1.0
AudienceValidator evaluation during TokenValidation doesn't honor ValidateAudience flag - We're upgrading from an older version 5.6.0 to 6.8.0. Seems like in the newest version, there was a change made to have AudienceValidator throw an exception here: https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/pull/36/commits/70c23ef985ad7d5c20a8e609b6c329642f311786 Which is fine and all. However, now it seems like if you specify a value for AudienceValidator AND set ValidateAudience = false, ValidateAudience does not get honored and we still evaluate the audience validator delegate rather than honoring the ValidateAudience flag. Seems like ValidateAudience needs to be accounted for. If it's set to false, we shouldn't throw on the delegate? We've worked around it for now by just manually nulling out AudienceValidator but still seems like a bug that should be fixed. Sample code: TokenValidationParameters tokenValidationParams = new TokenValidationParameters { AudienceValidator = AudienceValidator, }; tokenValidationParams.ValidateAudience = false; JwtSecurityTokenHandler.ValidateToken( token, tokenValidationParams, out validatedToken); Previous behavior: Passes validation New behavior: Throws Exception: Microsoft.Online.Common.Authentication.SecurityTokenValidation.Exceptions.MalformedTokenException: Malformed access token ---> Microsoft.IdentityModel.Tokens.SecurityTokenInvalidAudienceException: IDX10231: Audience validation failed. Delegate returned false, securitytoken: 'System.IdentityModel.Tokens.Jwt.JwtSecurityToken'.
non_defect
audiencevalidator evaluation during tokenvalidation doesn t honor validateaudience flag we re upgrading from an older version to seems like in the newest version there was a change made to have audiencevalidator throw an exception here which is fine and all however now it seems like if you specify a value for audiencevalidator and set validateaudience false validateaudience does not get honored and we still evaluate the audience validator delegate rather than honoring the validateaudience flag seems like validateaudience needs to be accounted for if it s set to false we shouldn t throw on the delegate we ve worked around it for now by just manually nulling out audiencevalidator but still seems like a bug that should be fixed sample code tokenvalidationparameters tokenvalidationparams new tokenvalidationparameters audiencevalidator audiencevalidator tokenvalidationparams validateaudience false jwtsecuritytokenhandler validatetoken token tokenvalidationparams out validatedtoken previous behavior passes validation new behavior throws exception microsoft online common authentication securitytokenvalidation exceptions malformedtokenexception malformed access token microsoft identitymodel tokens securitytokeninvalidaudienceexception audience validation failed delegate returned false securitytoken system identitymodel tokens jwt jwtsecuritytoken
0
58,915
16,918,902,903
IssuesEvent
2021-06-25 00:21:58
SAP/fundamental-ngx
https://api.github.com/repos/SAP/fundamental-ngx
opened
Notification Group is not focusable
Defect Hunting bug core denoland
#### Is this a bug, enhancement, or feature request? bug #### Briefly describe your proposal. no tab focus support <img width="415" alt="Screen Shot 2021-05-19 at 2 36 42 PM" src="https://user-images.githubusercontent.com/2471874/118880892-a3b0e980-b8af-11eb-9db5-1335abe37106.png">
1.0
Notification Group is not focusable - #### Is this a bug, enhancement, or feature request? bug #### Briefly describe your proposal. no tab focus support <img width="415" alt="Screen Shot 2021-05-19 at 2 36 42 PM" src="https://user-images.githubusercontent.com/2471874/118880892-a3b0e980-b8af-11eb-9db5-1335abe37106.png">
defect
notification group is not focusable is this a bug enhancement or feature request bug briefly describe your proposal no tab focus support img width alt screen shot at pm src
1
106,131
4,263,514,086
IssuesEvent
2016-07-12 01:10:40
Tributerre/LivrETS-Web
https://api.github.com/repos/Tributerre/LivrETS-Web
closed
Changer un titre dans la vue de cueillette
change priority: low
À la deuxième étape, changer le titre «Articles vendus par le vendeur» pour «Articles par le vendeur».
1.0
Changer un titre dans la vue de cueillette - À la deuxième étape, changer le titre «Articles vendus par le vendeur» pour «Articles par le vendeur».
non_defect
changer un titre dans la vue de cueillette à la deuxième étape changer le titre «articles vendus par le vendeur» pour «articles par le vendeur»
0
19,981
3,284,327,314
IssuesEvent
2015-10-28 16:13:13
wschoenell/chimera_imported_googlecode
https://api.github.com/repos/wschoenell/chimera_imported_googlecode
closed
Imageserver cannot autoload many images
auto-migrated Performance Priority-Medium Type-Defect Usability
``` If too many images are ready to be autoloaded, imageserver will fail due to Pyro having trouble registering the connections. More information is needed about this bug. ``` Original issue reported on code.google.com by `laughing...@gmail.com` on 7 Aug 2008 at 7:06
1.0
Imageserver cannot autoload many images - ``` If too many images are ready to be autoloaded, imageserver will fail due to Pyro having trouble registering the connections. More information is needed about this bug. ``` Original issue reported on code.google.com by `laughing...@gmail.com` on 7 Aug 2008 at 7:06
defect
imageserver cannot autoload many images if too many images are ready to be autoloaded imageserver will fail due to pyro having trouble registering the connections more information is needed about this bug original issue reported on code google com by laughing gmail com on aug at
1
241,170
26,256,674,846
IssuesEvent
2023-01-06 01:47:15
AlexRogalskiy/spark-patterns
https://api.github.com/repos/AlexRogalskiy/spark-patterns
opened
CVE-2022-24999 (High) detected in qs-6.10.1.tgz
security vulnerability
## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-path/node_modules/qs/package.json,/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - release-it-14.11.8.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.5.tgz - parse-url-6.0.0.tgz - parse-path-4.0.3.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark-patterns/commit/a53875b300612534ea9e204c7b863f1713ddcd1b">a53875b300612534ea9e204c7b863f1713ddcd1b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (release-it): 14.12.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24999 (High) detected in qs-6.10.1.tgz - ## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-path/node_modules/qs/package.json,/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - release-it-14.11.8.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.5.tgz - parse-url-6.0.0.tgz - parse-path-4.0.3.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark-patterns/commit/a53875b300612534ea9e204c7b863f1713ddcd1b">a53875b300612534ea9e204c7b863f1713ddcd1b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (release-it): 14.12.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules parse path node modules qs package json node modules qs package json dependency hierarchy release it tgz root library git url parse tgz git up tgz parse url tgz parse path tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs direct dependency fix resolution release it step up your open source security game with mend
0
256,245
8,127,087,587
IssuesEvent
2018-08-17 06:32:49
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
test suite: transform results.json into text summary
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
-----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1176 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: test suite: transform results.json into text summary Assigned to: Cyrus Harrison Category: Target version: Author: Cyrus Harrison Start: 08/23/2012 Due date: % Done: 0 Estimated time: Created: 08/23/2012 12:17 pm Updated: 05/02/2013 02:50 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Comments: Please add information about individual tests success/failure, not just success failure of a test file. Hi Everyone,This update provides several improvements to the json results recorded in by our test suite, including:- "results.json" now includes test suite params, all test script and cases results in one tree structure.- Proper locking for html index, and results.json for the multi-processing run case- visit_test_reports.py can generate a text report (all cases, or errors only) from a results.json file- visit_test_reports.py can email a report (all cases, or errors only) from a results.json fileThis check in resolves #1176, although some may want the text reports to be a little more readable/user friendly.This check in also enables email functionality necessary for a set of focused test suite runs.I also experimented with dynamically rendering the json results using only client-side html/javascript/jquery. This would be an alternativeto generating html in the test suite python code, a possible way to help alleviate some of our inode issues.The proof of concept worked, however the results are a little too big (~2MB) for us to expect good performance.To make sure this tinkering wasn't lost, I checked in these prototype pages under test/report_templates/.Committing to svn+ssh://cyrush@portal-auth.nersc.gov/project/projectdirs/visit/svn/visit/trunk/test ... A report_templates/js/purl.js A report_templates/prototype_json_index.html A report_templates/prototype_json_script_details.html M visit_test_common.py M visit_test_main.py M visit_test_reports.py M visit_test_suite.pyCommitted r20137-Cyrus This was resolved w/ r20137 as stated above.
1.0
test suite: transform results.json into text summary - -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1176 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: test suite: transform results.json into text summary Assigned to: Cyrus Harrison Category: Target version: Author: Cyrus Harrison Start: 08/23/2012 Due date: % Done: 0 Estimated time: Created: 08/23/2012 12:17 pm Updated: 05/02/2013 02:50 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Comments: Please add information about individual tests success/failure, not just success failure of a test file. Hi Everyone,This update provides several improvements to the json results recorded in by our test suite, including:- "results.json" now includes test suite params, all test script and cases results in one tree structure.- Proper locking for html index, and results.json for the multi-processing run case- visit_test_reports.py can generate a text report (all cases, or errors only) from a results.json file- visit_test_reports.py can email a report (all cases, or errors only) from a results.json fileThis check in resolves #1176, although some may want the text reports to be a little more readable/user friendly.This check in also enables email functionality necessary for a set of focused test suite runs.I also experimented with dynamically rendering the json results using only client-side html/javascript/jquery. This would be an alternativeto generating html in the test suite python code, a possible way to help alleviate some of our inode issues.The proof of concept worked, however the results are a little too big (~2MB) for us to expect good performance.To make sure this tinkering wasn't lost, I checked in these prototype pages under test/report_templates/.Committing to svn+ssh://cyrush@portal-auth.nersc.gov/project/projectdirs/visit/svn/visit/trunk/test ... A report_templates/js/purl.js A report_templates/prototype_json_index.html A report_templates/prototype_json_script_details.html M visit_test_common.py M visit_test_main.py M visit_test_reports.py M visit_test_suite.pyCommitted r20137-Cyrus This was resolved w/ r20137 as stated above.
non_defect
test suite transform results json into text summary redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject test suite transform results json into text summary assigned to cyrus harrison category target version author cyrus harrison start due date done estimated time created pm updated pm likelihood severity found in version impact medium expected use occasional os all support group any description comments please add information about individual tests success failure not just success failure of a test file hi everyone this update provides several improvements to the json results recorded in by our test suite including results json now includes test suite params all test script and cases results in one tree structure proper locking for html index and results json for the multi processing run case visit test reports py can generate a text report all cases or errors only from a results json file visit test reports py can email a report all cases or errors only from a results json filethis check in resolves although some may want the text reports to be a little more readable user friendly this check in also enables email functionality necessary for a set of focused test suite runs i also experimented with dynamically rendering the json results using only client side html javascript jquery this would be an alternativeto generating html in the test suite python code a possible way to help alleviate some of our inode issues the proof of concept worked however the results are a little too big for us to expect good performance to make sure this tinkering wasn t lost i checked in these prototype pages under test report templates committing to svn ssh cyrush portal auth nersc gov project projectdirs visit svn visit trunk test a report templates js purl js a report templates prototype json index html a report templates prototype json script details html m visit test common py m visit test main py m visit test reports py m visit test suite pycommitted cyrus this was resolved w as stated above
0
224,786
17,775,044,475
IssuesEvent
2021-08-30 18:04:55
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: sysbench/oltp_point_select/nodes=3/cpu=32/conc=128 failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-queries
roachtest.sysbench/oltp_point_select/nodes=3/cpu=32/conc=128 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3377334&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3377334&tab=artifacts#/sysbench/oltp_point_select/nodes=3/cpu=32/conc=128) on master @ [6700cf65e82a029a1ff76a75250c709aa9637170](https://github.com/cockroachdb/cockroach/commits/6700cf65e82a029a1ff76a75250c709aa9637170): ``` Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString cluster.go:1249,context.go:89,cluster.go:1237,test_runner.go:866: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3377334-1630217394-83-n4cpu32 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 2: dead (exit status 134) 3: dead (exit status 134) 1: 11896 Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 134) (1) UNCLASSIFIED_PROBLEM Wraps: (2) secondary error attachment | 3: dead (exit status 134) | (1) attached stack trace | -- stack trace: | | main.glob..func14 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173 | | main.wrap.func1 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281 | | github.com/spf13/cobra.(*Command).execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856 | | github.com/spf13/cobra.(*Command).ExecuteC | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960 | | github.com/spf13/cobra.(*Command).Execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897 | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107 | | runtime.main | | /usr/local/go/src/runtime/proc.go:225 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1371 | Wraps: (2) 3: dead (exit status 134) | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (3) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107 | runtime.main | /usr/local/go/src/runtime/proc.go:225 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1371 Wraps: (4) 2: dead (exit status 134) Error types: (1) errors.Unclassified (2) *secondary.withSecondaryError (3) *withstack.withStack (4) *errutil.leafError ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sysbench/oltp_point_select/nodes=3/cpu=32/conc=128.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: sysbench/oltp_point_select/nodes=3/cpu=32/conc=128 failed - roachtest.sysbench/oltp_point_select/nodes=3/cpu=32/conc=128 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3377334&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3377334&tab=artifacts#/sysbench/oltp_point_select/nodes=3/cpu=32/conc=128) on master @ [6700cf65e82a029a1ff76a75250c709aa9637170](https://github.com/cockroachdb/cockroach/commits/6700cf65e82a029a1ff76a75250c709aa9637170): ``` Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString cluster.go:1249,context.go:89,cluster.go:1237,test_runner.go:866: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3377334-1630217394-83-n4cpu32 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 2: dead (exit status 134) 3: dead (exit status 134) 1: 11896 Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 134) (1) UNCLASSIFIED_PROBLEM Wraps: (2) secondary error attachment | 3: dead (exit status 134) | (1) attached stack trace | -- stack trace: | | main.glob..func14 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173 | | main.wrap.func1 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281 | | github.com/spf13/cobra.(*Command).execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856 | | github.com/spf13/cobra.(*Command).ExecuteC | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960 | | github.com/spf13/cobra.(*Command).Execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897 | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107 | | runtime.main | | /usr/local/go/src/runtime/proc.go:225 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1371 | Wraps: (2) 3: dead (exit status 134) | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (3) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107 | runtime.main | /usr/local/go/src/runtime/proc.go:225 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1371 Wraps: (4) 2: dead (exit status 134) Error types: (1) errors.Unclassified (2) *secondary.withSecondaryError (3) *withstack.withStack (4) *errutil.leafError ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sysbench/oltp_point_select/nodes=3/cpu=32/conc=128.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_defect
roachtest sysbench oltp point select nodes cpu conc failed roachtest sysbench oltp point select nodes cpu conc with on master error types withstack withstack errutil withprefix errors errorstring cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead exit status dead exit status error unclassified problem dead exit status unclassified problem wraps secondary error attachment dead exit status attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead exit status error types withstack withstack errutil leaferror wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead exit status error types errors unclassified secondary withsecondaryerror withstack withstack errutil leaferror reproduce see cc cockroachdb kv triage
0
5,013
2,610,164,373
IssuesEvent
2015-02-26 18:52:07
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Gameplay Error
auto-migrated Priority-Medium Type-Defect
``` Jango Fett costs 20credits and takes 0,5 sec to build ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 1:09
1.0
Gameplay Error - ``` Jango Fett costs 20credits and takes 0,5 sec to build ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 1:09
defect
gameplay error jango fett costs and takes sec to build original issue reported on code google com by gmail com on may at
1
120,884
12,099,585,229
IssuesEvent
2020-04-20 12:26:35
davehenryjones/WellbeingJam2020
https://api.github.com/repos/davehenryjones/WellbeingJam2020
closed
Create a data schema and example
documentation enhancement
- Data standard - Example data sets available for download - Naming conventions / date convention
1.0
Create a data schema and example - - Data standard - Example data sets available for download - Naming conventions / date convention
non_defect
create a data schema and example data standard example data sets available for download naming conventions date convention
0
59,245
17,016,669,124
IssuesEvent
2021-07-02 13:02:04
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
not all the addresses are correctly displayed
Component: nominatim Priority: major Type: defect
**[Submitted to the original trac issue database at 8.34am, Friday, 4th September 2015]** Good day. I am having the following problem - not all the addresses are correctly displayed. At the moment the problem is observed only in Chile. Here is an example query: http://map.sinopacificukraine.com/nomin sdetails=1 openstreetmap map api(working): http://nominatim.openstreetmap.org/reve sdetails=1
1.0
not all the addresses are correctly displayed - **[Submitted to the original trac issue database at 8.34am, Friday, 4th September 2015]** Good day. I am having the following problem - not all the addresses are correctly displayed. At the moment the problem is observed only in Chile. Here is an example query: http://map.sinopacificukraine.com/nomin sdetails=1 openstreetmap map api(working): http://nominatim.openstreetmap.org/reve sdetails=1
defect
not all the addresses are correctly displayed good day i am having the following problem not all the addresses are correctly displayed at the moment the problem is observed only in chile here is an example query sdetails openstreetmap map api working sdetails
1
19,235
3,162,158,852
IssuesEvent
2015-09-19 02:13:12
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
dart2js: NoSuchMethodError prints interceptor of receiver instead of receiver
Area-Dart2JS Priority-Unassigned Triaged Type-Defect
Following program prints: ----- NoSuchMethodError : method not found: 'add' Receiver: Instance of 'JSInt' Arguments: [88] ----- Should print: ----- NoSuchMethodError : method not found: 'add' Receiver: 666 Arguments: [88] ----- The problem is that the nsm stub uses 'this'. In the general case it needs to use '$receiver' when 'this' is an interceptor. (I think this is as simple as passing $receiver instead of 'this', but consider what the dummy receiver optimization looks like for classes implementing/not implementing noSuchMethod). &nbsp;&nbsp;&nbsp;&nbsp;add$1: function($receiver, $0) { &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;return this.noSuchMethod$1(this, H.createInvocationMirror(&quot;add&quot;, &quot;add$1&quot;, 0, [$0], [])); &nbsp;&nbsp;&nbsp;&nbsp;}, -------- class N { &nbsp;&nbsp;noSuchMethod(i) =&gt; 42; } class CCCC { &nbsp;&nbsp;foo([a, b]) =&gt; '[$a, $b]'; } get NEVER =&gt; new DateTime.now().millisecondsSinceEpoch &lt; 2; var g; main() { &nbsp;&nbsp;var c = new N(); &nbsp;&nbsp;if (!NEVER) c = 666; &nbsp;&nbsp;try { &nbsp;&nbsp;&nbsp;&nbsp;g = c..toString()..add(88); &nbsp;&nbsp;} catch (e) { &nbsp;&nbsp;&nbsp;&nbsp;print(e); &nbsp;&nbsp;} }
1.0
dart2js: NoSuchMethodError prints interceptor of receiver instead of receiver - Following program prints: ----- NoSuchMethodError : method not found: 'add' Receiver: Instance of 'JSInt' Arguments: [88] ----- Should print: ----- NoSuchMethodError : method not found: 'add' Receiver: 666 Arguments: [88] ----- The problem is that the nsm stub uses 'this'. In the general case it needs to use '$receiver' when 'this' is an interceptor. (I think this is as simple as passing $receiver instead of 'this', but consider what the dummy receiver optimization looks like for classes implementing/not implementing noSuchMethod). &nbsp;&nbsp;&nbsp;&nbsp;add$1: function($receiver, $0) { &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;return this.noSuchMethod$1(this, H.createInvocationMirror(&quot;add&quot;, &quot;add$1&quot;, 0, [$0], [])); &nbsp;&nbsp;&nbsp;&nbsp;}, -------- class N { &nbsp;&nbsp;noSuchMethod(i) =&gt; 42; } class CCCC { &nbsp;&nbsp;foo([a, b]) =&gt; '[$a, $b]'; } get NEVER =&gt; new DateTime.now().millisecondsSinceEpoch &lt; 2; var g; main() { &nbsp;&nbsp;var c = new N(); &nbsp;&nbsp;if (!NEVER) c = 666; &nbsp;&nbsp;try { &nbsp;&nbsp;&nbsp;&nbsp;g = c..toString()..add(88); &nbsp;&nbsp;} catch (e) { &nbsp;&nbsp;&nbsp;&nbsp;print(e); &nbsp;&nbsp;} }
defect
nosuchmethoderror prints interceptor of receiver instead of receiver following program prints nosuchmethoderror method not found add receiver instance of jsint arguments should print nosuchmethoderror method not found add receiver arguments the problem is that the nsm stub uses this in the general case it needs to use receiver when this is an interceptor i think this is as simple as passing receiver instead of this but consider what the dummy receiver optimization looks like for classes implementing not implementing nosuchmethod nbsp nbsp nbsp nbsp add function receiver nbsp nbsp nbsp nbsp nbsp nbsp return this nosuchmethod this h createinvocationmirror quot add quot quot add quot nbsp nbsp nbsp nbsp class n nbsp nbsp nosuchmethod i gt class cccc nbsp nbsp foo gt get never gt new datetime now millisecondssinceepoch lt var g main nbsp nbsp var c new n nbsp nbsp if never c nbsp nbsp try nbsp nbsp nbsp nbsp g c tostring add nbsp nbsp catch e nbsp nbsp nbsp nbsp print e nbsp nbsp
1
27,519
5,038,857,904
IssuesEvent
2016-12-18 14:02:21
chandanbansal/gmail-backup
https://api.github.com/repos/chandanbansal/gmail-backup
opened
Latest Downloads not available
auto-migrated Priority-Medium Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Saturday Apr 23, 2016 at 10:31 GMT_ _Originally opened as https://github.com/terry10/gmail-backup-com/issues/12_ ---- ``` What steps will reproduce the problem? 1. Use the "Check new versions" in the "Help" menu 2. Lists version 20 available at http://code.google.com/p/gmail-backup-com/downloads/list 3. Follow link; latest version listed is 15. What is the expected output? What do you see instead? Expect to see version 20; only see v15 What version of the product are you using? On what operating system? v15, Windows 7 Please provide any additional information below. ``` Original issue reported on code.google.com by `rpeter...@teraverde.com` on 31 Aug 2012 at 2:47
1.0
Latest Downloads not available - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Saturday Apr 23, 2016 at 10:31 GMT_ _Originally opened as https://github.com/terry10/gmail-backup-com/issues/12_ ---- ``` What steps will reproduce the problem? 1. Use the "Check new versions" in the "Help" menu 2. Lists version 20 available at http://code.google.com/p/gmail-backup-com/downloads/list 3. Follow link; latest version listed is 15. What is the expected output? What do you see instead? Expect to see version 20; only see v15 What version of the product are you using? On what operating system? v15, Windows 7 Please provide any additional information below. ``` Original issue reported on code.google.com by `rpeter...@teraverde.com` on 31 Aug 2012 at 2:47
defect
latest downloads not available issue by saturday apr at gmt originally opened as what steps will reproduce the problem use the check new versions in the help menu lists version available at follow link latest version listed is what is the expected output what do you see instead expect to see version only see what version of the product are you using on what operating system windows please provide any additional information below original issue reported on code google com by rpeter teraverde com on aug at
1
158,710
13,746,081,896
IssuesEvent
2020-10-06 04:46:50
edgexfoundry/edgex-go
https://api.github.com/repos/edgexfoundry/edgex-go
closed
Add Core Metadata v2 API Response Examples
bug documentation
<!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 📚 Docs or Wiki Bug Report ### Description <!-- ✍️edit:--> Example data shows data type currently, needs to update it to a valid value. ## 🔬 Minimal Reproduction ### What's the affected URL?** <!-- ✍️edit:--> https://app.swaggerhub.com/apis-docs/EdgeXFoundry1/core-metadata/2.x ### Expected vs Actual Behavior** <!-- If applicable please describe the difference between the expected and actual behavior after following the steps to reproduce. --> <!-- ✍️edit:--> ## 📷Screenshot <!-- Often a screenshot can help to capture the issue better than a long description. --> <!-- ✍️upload a screenshot:--> ## 🔥 Exception or Error <pre><code> <!-- If the issue is accompanied by an exception or an error, please share it below: --> <!-- ✍️--> </code></pre> ## 🌍 Your Environment <!-- How are you viewing the docs? What browser? What version? : --> <!-- ✍️-->
1.0
Add Core Metadata v2 API Response Examples - <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 📚 Docs or Wiki Bug Report ### Description <!-- ✍️edit:--> Example data shows data type currently, needs to update it to a valid value. ## 🔬 Minimal Reproduction ### What's the affected URL?** <!-- ✍️edit:--> https://app.swaggerhub.com/apis-docs/EdgeXFoundry1/core-metadata/2.x ### Expected vs Actual Behavior** <!-- If applicable please describe the difference between the expected and actual behavior after following the steps to reproduce. --> <!-- ✍️edit:--> ## 📷Screenshot <!-- Often a screenshot can help to capture the issue better than a long description. --> <!-- ✍️upload a screenshot:--> ## 🔥 Exception or Error <pre><code> <!-- If the issue is accompanied by an exception or an error, please share it below: --> <!-- ✍️--> </code></pre> ## 🌍 Your Environment <!-- How are you viewing the docs? What browser? What version? : --> <!-- ✍️-->
non_defect
add core metadata api response examples 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 oh hi there 😄 to expedite issue processing please search open and closed issues before submitting a new one existing issues often contain information about workarounds resolution or progress updates 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 📚 docs or wiki bug report description example data shows data type currently needs to update it to a valid value 🔬 minimal reproduction what s the affected url expected vs actual behavior 📷screenshot 🔥 exception or error 🌍 your environment
0
78,354
27,451,695,397
IssuesEvent
2023-03-02 17:52:12
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Element-web sends, and responds to, hopeless `m.room_key_request` messages
T-Defect S-Tolerable A-E2EE O-Frequent Team: Crypto
The mitigations to CVE-2022-39249 and friends mean that we will never trust an `m.forwarded_room_key` message from another user's device. Accordingly, there is no point sending `m.room_key_request` messages to other users. Nevertheless we still send such messages, and indeed respond to them. Not doing so would reduce the volume of to-device messages that need to be sent around and processed, and might also make debugging UTD errors easier because of the reduced noise.
1.0
Element-web sends, and responds to, hopeless `m.room_key_request` messages - The mitigations to CVE-2022-39249 and friends mean that we will never trust an `m.forwarded_room_key` message from another user's device. Accordingly, there is no point sending `m.room_key_request` messages to other users. Nevertheless we still send such messages, and indeed respond to them. Not doing so would reduce the volume of to-device messages that need to be sent around and processed, and might also make debugging UTD errors easier because of the reduced noise.
defect
element web sends and responds to hopeless m room key request messages the mitigations to cve and friends mean that we will never trust an m forwarded room key message from another user s device accordingly there is no point sending m room key request messages to other users nevertheless we still send such messages and indeed respond to them not doing so would reduce the volume of to device messages that need to be sent around and processed and might also make debugging utd errors easier because of the reduced noise
1
352,592
25,073,993,020
IssuesEvent
2022-11-07 14:15:44
eclipse/cloe
https://api.github.com/repos/eclipse/cloe
opened
Add instructions to install cloe-launch without cloning
documentation
It can be useful for users who use an Artifactory for package management to just install the `cloe-launch` Python tool. This can be done with: pip install 'git+https://github.com/eclipse/cloe@master#subdirectory=cli' We should document this.
1.0
Add instructions to install cloe-launch without cloning - It can be useful for users who use an Artifactory for package management to just install the `cloe-launch` Python tool. This can be done with: pip install 'git+https://github.com/eclipse/cloe@master#subdirectory=cli' We should document this.
non_defect
add instructions to install cloe launch without cloning it can be useful for users who use an artifactory for package management to just install the cloe launch python tool this can be done with pip install git we should document this
0
60,001
8,393,952,379
IssuesEvent
2018-10-09 22:14:57
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
reopened
"%NIM%/config/nim.cfg" is not being read
Documentation
`nim compile ...` is not reading the configuration file at "%NIM%/config/nim.cfg" (as it should be per the [documentation](https://nim-lang.org/docs/nimc.html#compiler-usage-configuration-files)). ```shell C:>ver Microsoft Windows [Version 10.0.17134.320] C:>nim --version Nim Compiler Version 0.19.0 [Windows: amd64] Compiled at 2018-09-26 Copyright (c) 2006-2018 by Andreas Rumpf git hash: f6c5c636bb1a1f4e1301ae0ba5a8afecef439132 active boot switches: -d:release ```
1.0
"%NIM%/config/nim.cfg" is not being read - `nim compile ...` is not reading the configuration file at "%NIM%/config/nim.cfg" (as it should be per the [documentation](https://nim-lang.org/docs/nimc.html#compiler-usage-configuration-files)). ```shell C:>ver Microsoft Windows [Version 10.0.17134.320] C:>nim --version Nim Compiler Version 0.19.0 [Windows: amd64] Compiled at 2018-09-26 Copyright (c) 2006-2018 by Andreas Rumpf git hash: f6c5c636bb1a1f4e1301ae0ba5a8afecef439132 active boot switches: -d:release ```
non_defect
nim config nim cfg is not being read nim compile is not reading the configuration file at nim config nim cfg as it should be per the shell c ver microsoft windows c nim version nim compiler version compiled at copyright c by andreas rumpf git hash active boot switches d release
0
33,729
7,208,001,921
IssuesEvent
2018-02-07 00:41:01
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Shift + Ctrl = Death
C: Peacock P: minor T: defect
## Rationale If you have a VTK window selected in Peacock and you press Shift+Ctrl together at the same time it crashes with: ``` Traceback (most recent call last): File "/opt/moose/miniconda/lib/python2.7/site-packages/vtk/qt/QVTKRenderWindowInteractor.py", line 440, in keyPressEvent if shift and len(keySym) == 1 and keySym.isalpha(): TypeError: object of type 'NoneType' has no len() ``` ## Description Good luck! ## Impact Technically a bug - but definitely not important!
1.0
Shift + Ctrl = Death - ## Rationale If you have a VTK window selected in Peacock and you press Shift+Ctrl together at the same time it crashes with: ``` Traceback (most recent call last): File "/opt/moose/miniconda/lib/python2.7/site-packages/vtk/qt/QVTKRenderWindowInteractor.py", line 440, in keyPressEvent if shift and len(keySym) == 1 and keySym.isalpha(): TypeError: object of type 'NoneType' has no len() ``` ## Description Good luck! ## Impact Technically a bug - but definitely not important!
defect
shift ctrl death rationale if you have a vtk window selected in peacock and you press shift ctrl together at the same time it crashes with traceback most recent call last file opt moose miniconda lib site packages vtk qt qvtkrenderwindowinteractor py line in keypressevent if shift and len keysym and keysym isalpha typeerror object of type nonetype has no len description good luck impact technically a bug but definitely not important
1
36,908
8,188,212,799
IssuesEvent
2018-08-30 00:29:13
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
CFDI de pago en web - no está enviando el número de parcialidad cuando metPago=PPD
bug defect
Cuando el método de pago "MetodoDePagoDR" del documento relacionado (factura que se está pagando) es PPD, se debe informar el número de parcialidad. Te mandé un correo explicando el tema
1.0
CFDI de pago en web - no está enviando el número de parcialidad cuando metPago=PPD - Cuando el método de pago "MetodoDePagoDR" del documento relacionado (factura que se está pagando) es PPD, se debe informar el número de parcialidad. Te mandé un correo explicando el tema
defect
cfdi de pago en web no está enviando el número de parcialidad cuando metpago ppd cuando el método de pago metododepagodr del documento relacionado factura que se está pagando es ppd se debe informar el número de parcialidad te mandé un correo explicando el tema
1
10,034
2,618,932,355
IssuesEvent
2015-03-03 00:00:47
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
destroyer v1 can fire bombs
auto-migrated Component-Logic Component-UI Priority-Medium Type-Defect
``` Game version: 0.95.125 Operating System: Windows XP 32-bit Java runtime version: jdk1.7.0_04 Installed using the Launcher? yes Game language (en, hu, de): de What steps will reproduce the problem? 1. In fleet-equipment screen, have the lasers slot of a destroyer v1 selected. Be producing bomb v1 in production screen. 2. Open the quick production panel and increase number of bombs v1 to produce by 1. This lets the +1 button for the lasers show up, even though I have a full set of lasers there. 3. Clicking on the button adds bombs v1 until the destroyer has 6 bombs. I also checked that I can actually fire the bomb and damage enemy defense. It is also possible to equip a cruiser 1 with rockets but didn't fire them. So probably would work for all kinds of illegal combinations. In the save, fleet is at Achilles ready to reproduce. ``` Original issue reported on code.google.com by `anonymis...@arcor.de` on 23 Jan 2013 at 7:13 Attachments: * [save-2013-01-23-09-09-05-687.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-712/comment-0/save-2013-01-23-09-09-05-687.xml.gz) * [info-2013-01-23-09-09-05-687.xml](https://storage.googleapis.com/google-code-attachments/open-ig/issue-712/comment-0/info-2013-01-23-09-09-05-687.xml)
1.0
destroyer v1 can fire bombs - ``` Game version: 0.95.125 Operating System: Windows XP 32-bit Java runtime version: jdk1.7.0_04 Installed using the Launcher? yes Game language (en, hu, de): de What steps will reproduce the problem? 1. In fleet-equipment screen, have the lasers slot of a destroyer v1 selected. Be producing bomb v1 in production screen. 2. Open the quick production panel and increase number of bombs v1 to produce by 1. This lets the +1 button for the lasers show up, even though I have a full set of lasers there. 3. Clicking on the button adds bombs v1 until the destroyer has 6 bombs. I also checked that I can actually fire the bomb and damage enemy defense. It is also possible to equip a cruiser 1 with rockets but didn't fire them. So probably would work for all kinds of illegal combinations. In the save, fleet is at Achilles ready to reproduce. ``` Original issue reported on code.google.com by `anonymis...@arcor.de` on 23 Jan 2013 at 7:13 Attachments: * [save-2013-01-23-09-09-05-687.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-712/comment-0/save-2013-01-23-09-09-05-687.xml.gz) * [info-2013-01-23-09-09-05-687.xml](https://storage.googleapis.com/google-code-attachments/open-ig/issue-712/comment-0/info-2013-01-23-09-09-05-687.xml)
defect
destroyer can fire bombs game version operating system windows xp bit java runtime version installed using the launcher yes game language en hu de de what steps will reproduce the problem in fleet equipment screen have the lasers slot of a destroyer selected be producing bomb in production screen open the quick production panel and increase number of bombs to produce by this lets the button for the lasers show up even though i have a full set of lasers there clicking on the button adds bombs until the destroyer has bombs i also checked that i can actually fire the bomb and damage enemy defense it is also possible to equip a cruiser with rockets but didn t fire them so probably would work for all kinds of illegal combinations in the save fleet is at achilles ready to reproduce original issue reported on code google com by anonymis arcor de on jan at attachments
1
29,816
24,302,446,240
IssuesEvent
2022-09-29 14:47:26
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
Reduce Cloud Spend 💸
infrastructure Priority: P3 CAP stale
We should take inventory of our team's resources in Azure + GCP and cleanup/reduce wherever possible. - [ ] Create a notion doc cataloguing CAP's existing infra - [ ] Figure out how much we're currently spending on resources - [ ] Identify areas to reduce spend (ex. use single Postgres instance for ODIS Alfajores signers) - [ ] Measure cost reduction - [ ] Optional: identify ways to keep costs down going forward (ex. cost alerting) - [ ] Optional: share with larger team to help other teams reduce costs
1.0
Reduce Cloud Spend 💸 - We should take inventory of our team's resources in Azure + GCP and cleanup/reduce wherever possible. - [ ] Create a notion doc cataloguing CAP's existing infra - [ ] Figure out how much we're currently spending on resources - [ ] Identify areas to reduce spend (ex. use single Postgres instance for ODIS Alfajores signers) - [ ] Measure cost reduction - [ ] Optional: identify ways to keep costs down going forward (ex. cost alerting) - [ ] Optional: share with larger team to help other teams reduce costs
non_defect
reduce cloud spend 💸 we should take inventory of our team s resources in azure gcp and cleanup reduce wherever possible create a notion doc cataloguing cap s existing infra figure out how much we re currently spending on resources identify areas to reduce spend ex use single postgres instance for odis alfajores signers measure cost reduction optional identify ways to keep costs down going forward ex cost alerting optional share with larger team to help other teams reduce costs
0
48,762
13,184,732,200
IssuesEvent
2020-08-12 19:59:38
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
ports fail to compile on ubuntu 8.10 x86_64 (Trac #154)
Incomplete Migration Migrated from Trac defect tools/ports
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/154 , reported by klas and owned by cgils_</summary> <p> ```json { "status": "closed", "changetime": "2009-03-06T01:48:15", "description": " Running make to compile the ports source, I get\n\n....\n\nmake[2]: Leaving directory `/a/klas/icecube/port_source/src/package1.0'\n===> making all in src/pextlib1.0\nmake[2]: Entering directory `/a/klas/icecube/port_source/src/pextlib1.0'\ngcc -c -DUSE_TCL_STUBS -W -Wall -pedantic -g -O2 -DHAVE_CONFIG_H -I.. -I. -I\"/usr/include/tcl8.4\" -DHAVE_LIMITS_H=1 -DHAVE_UNISTD_H=1 -DHAVE_SYS_PARAM_H=1 -DUSE_THREAD_ALLOC=1 -D_REENTRANT=1 -D_THREAD_SAFE=1 -DHAVE_PTHREAD_ATTR_SETSTACKSIZE=1 -DHAVE_PTHREAD_ATFORK=1 -DTCL_THREADS=1 -DPEEK_XCLOSEIM=1 -D_LARGEFILE64_SOURCE=1 -DTCL_WIDE_INT_IS_LONG=1 -DHAVE_GETCWD=1 -DHAVE_OPENDIR=1 -DHAVE_STRSTR=1 -DHAVE_STRTOL=1 -DHAVE_STRTOLL=1 -DHAVE_STRTOULL=1 -DHAVE_TMPNAM=1 -DHAVE_WAITPID=1 -DHAVE_GETPWUID_R_5=1 -DHAVE_GETPWUID_R=1 -DHAVE_GETPWNAM_R_5=1 -DHAVE_GETPWNAM_R=1 -DHAVE_GETGRGID_R_5=1 -DHAVE_GETGRGID_R=1 -DHAVE_GETGRNAM_R_5=1 -DHAVE_GETGRNAM_R=1 -DHAVE_GETHOSTBYNAME_R_6=1 -DHAVE_GETHOSTBYNAME_R=1 -DHAVE_GETHOSTBYADDR_R_8=1 -DHAVE_GETHOSTBYADDR_R=1 -DUSE_TERMIOS=1 -DHAVE_SYS_TIME_H=1 -DTIME_WITH_SYS_TIME=1 -DHAVE_TM_ZONE=1 -DHAVE_GMTIME_R=1 -DHAVE_LOCALTIME_R=1 -DHAVE_TM_GMTOFF=1 -DHAVE_TIMEZONE_VAR=1 -DHAVE_ST_BLKSIZE=1 -DSTDC_HEADERS=1 -DHAVE_SIGNED_CHAR=1 -DHAVE_LANGINFO=1 -DHAVE_SYS_IOCTL_H=1 -fPIC xinstall.c -o xinstall.o\nxinstall.c: In function \u2018InstallCmd\u2019:\nxinstall.c:142: warning: unused variable \u2018flags\u2019\nxinstall.c: In function \u2018install\u2019:\nxinstall.c:422: warning: unused parameter \u2018fset\u2019\nxinstall.c: In function \u2018create_newfile\u2019:\nxinstall.c:856: warning: unused parameter \u2018sbp\u2019\nxinstall.c: In function \u2018trymmap\u2019:\nxinstall.c:1091: warning: unused parameter \u2018fd\u2019\nxinstall.c: In function \u2018InstallCmd\u2019:\nxinstall.c:344: warning: \u2018fset\u2019 may be used uninitialized in this function\nIn function \u2018open\u2019,\n inlined from \u2018create_newfile\u2019 at xinstall.c:900,\n inlined from \u2018install\u2019 at xinstall.c:524:make[2]: Leaving directory `/a/klas/icecube/port_source/src/package1.0'\n/usr/include/bits/fcntl2.h:45: error: call to \u2018__open_too_many_args\u2019 declared with attribute error: open can be called either with 2 or 3 arguments, not more\n/usr/include/bits/fcntl2.h:45: error: call to \u2018__open_too_many_args\u2019 declared with attribute error: open can be called either with 2 or 3 arguments, not more\nmake[2]: *** [xinstall.o] Error 1\nmake[2]: Leaving directory `/a/klas/icecube/port_source/src/pextlib1.0'\nmake[1]: *** [all] Error 1\nmake[1]: Leaving directory `/a/klas/icecube/port_source/src'\nmake: *** [all] Error 1", "reporter": "klas", "cc": "", "resolution": "fixed", "_ts": "1236304095000000", "component": "tools/ports", "summary": "ports fail to compile on ubuntu 8.10 x86_64", "priority": "normal", "keywords": "", "time": "2008-11-19T02:42:53", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
1.0
ports fail to compile on ubuntu 8.10 x86_64 (Trac #154) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/154 , reported by klas and owned by cgils_</summary> <p> ```json { "status": "closed", "changetime": "2009-03-06T01:48:15", "description": " Running make to compile the ports source, I get\n\n....\n\nmake[2]: Leaving directory `/a/klas/icecube/port_source/src/package1.0'\n===> making all in src/pextlib1.0\nmake[2]: Entering directory `/a/klas/icecube/port_source/src/pextlib1.0'\ngcc -c -DUSE_TCL_STUBS -W -Wall -pedantic -g -O2 -DHAVE_CONFIG_H -I.. -I. -I\"/usr/include/tcl8.4\" -DHAVE_LIMITS_H=1 -DHAVE_UNISTD_H=1 -DHAVE_SYS_PARAM_H=1 -DUSE_THREAD_ALLOC=1 -D_REENTRANT=1 -D_THREAD_SAFE=1 -DHAVE_PTHREAD_ATTR_SETSTACKSIZE=1 -DHAVE_PTHREAD_ATFORK=1 -DTCL_THREADS=1 -DPEEK_XCLOSEIM=1 -D_LARGEFILE64_SOURCE=1 -DTCL_WIDE_INT_IS_LONG=1 -DHAVE_GETCWD=1 -DHAVE_OPENDIR=1 -DHAVE_STRSTR=1 -DHAVE_STRTOL=1 -DHAVE_STRTOLL=1 -DHAVE_STRTOULL=1 -DHAVE_TMPNAM=1 -DHAVE_WAITPID=1 -DHAVE_GETPWUID_R_5=1 -DHAVE_GETPWUID_R=1 -DHAVE_GETPWNAM_R_5=1 -DHAVE_GETPWNAM_R=1 -DHAVE_GETGRGID_R_5=1 -DHAVE_GETGRGID_R=1 -DHAVE_GETGRNAM_R_5=1 -DHAVE_GETGRNAM_R=1 -DHAVE_GETHOSTBYNAME_R_6=1 -DHAVE_GETHOSTBYNAME_R=1 -DHAVE_GETHOSTBYADDR_R_8=1 -DHAVE_GETHOSTBYADDR_R=1 -DUSE_TERMIOS=1 -DHAVE_SYS_TIME_H=1 -DTIME_WITH_SYS_TIME=1 -DHAVE_TM_ZONE=1 -DHAVE_GMTIME_R=1 -DHAVE_LOCALTIME_R=1 -DHAVE_TM_GMTOFF=1 -DHAVE_TIMEZONE_VAR=1 -DHAVE_ST_BLKSIZE=1 -DSTDC_HEADERS=1 -DHAVE_SIGNED_CHAR=1 -DHAVE_LANGINFO=1 -DHAVE_SYS_IOCTL_H=1 -fPIC xinstall.c -o xinstall.o\nxinstall.c: In function \u2018InstallCmd\u2019:\nxinstall.c:142: warning: unused variable \u2018flags\u2019\nxinstall.c: In function \u2018install\u2019:\nxinstall.c:422: warning: unused parameter \u2018fset\u2019\nxinstall.c: In function \u2018create_newfile\u2019:\nxinstall.c:856: warning: unused parameter \u2018sbp\u2019\nxinstall.c: In function \u2018trymmap\u2019:\nxinstall.c:1091: warning: unused parameter \u2018fd\u2019\nxinstall.c: In function \u2018InstallCmd\u2019:\nxinstall.c:344: warning: \u2018fset\u2019 may be used uninitialized in this function\nIn function \u2018open\u2019,\n inlined from \u2018create_newfile\u2019 at xinstall.c:900,\n inlined from \u2018install\u2019 at xinstall.c:524:make[2]: Leaving directory `/a/klas/icecube/port_source/src/package1.0'\n/usr/include/bits/fcntl2.h:45: error: call to \u2018__open_too_many_args\u2019 declared with attribute error: open can be called either with 2 or 3 arguments, not more\n/usr/include/bits/fcntl2.h:45: error: call to \u2018__open_too_many_args\u2019 declared with attribute error: open can be called either with 2 or 3 arguments, not more\nmake[2]: *** [xinstall.o] Error 1\nmake[2]: Leaving directory `/a/klas/icecube/port_source/src/pextlib1.0'\nmake[1]: *** [all] Error 1\nmake[1]: Leaving directory `/a/klas/icecube/port_source/src'\nmake: *** [all] Error 1", "reporter": "klas", "cc": "", "resolution": "fixed", "_ts": "1236304095000000", "component": "tools/ports", "summary": "ports fail to compile on ubuntu 8.10 x86_64", "priority": "normal", "keywords": "", "time": "2008-11-19T02:42:53", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
defect
ports fail to compile on ubuntu trac migrated from reported by klas and owned by cgils json status closed changetime description running make to compile the ports source i get n n n nmake leaving directory a klas icecube port source src n making all in src nmake entering directory a klas icecube port source src ngcc c duse tcl stubs w wall pedantic g dhave config h i i i usr include dhave limits h dhave unistd h dhave sys param h duse thread alloc d reentrant d thread safe dhave pthread attr setstacksize dhave pthread atfork dtcl threads dpeek xcloseim d source dtcl wide int is long dhave getcwd dhave opendir dhave strstr dhave strtol dhave strtoll dhave strtoull dhave tmpnam dhave waitpid dhave getpwuid r dhave getpwuid r dhave getpwnam r dhave getpwnam r dhave getgrgid r dhave getgrgid r dhave getgrnam r dhave getgrnam r dhave gethostbyname r dhave gethostbyname r dhave gethostbyaddr r dhave gethostbyaddr r duse termios dhave sys time h dtime with sys time dhave tm zone dhave gmtime r dhave localtime r dhave tm gmtoff dhave timezone var dhave st blksize dstdc headers dhave signed char dhave langinfo dhave sys ioctl h fpic xinstall c o xinstall o nxinstall c in function nxinstall c warning unused variable nxinstall c in function nxinstall c warning unused parameter nxinstall c in function newfile nxinstall c warning unused parameter nxinstall c in function nxinstall c warning unused parameter nxinstall c in function nxinstall c warning may be used uninitialized in this function nin function n inlined from newfile at xinstall c n inlined from at xinstall c make leaving directory a klas icecube port source src n usr include bits h error call to open too many args declared with attribute error open can be called either with or arguments not more n usr include bits h error call to open too many args declared with attribute error open can be called either with or arguments not more nmake error nmake leaving directory a klas icecube port source src nmake error nmake leaving directory a klas icecube port source src nmake error reporter klas cc resolution fixed ts component tools ports summary ports fail to compile on ubuntu priority normal keywords time milestone owner cgils type defect
1
76,078
26,226,776,355
IssuesEvent
2023-01-04 19:28:40
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
Notification on iOS (wrong numbers of notifiaction in the red bubble)
T-Defect
### Steps to reproduce I have the problem that every message that the Element Messenger receives shows a three in the notification bubble instead of a one. I have now tried all possible notification options both in the iOS app and directly in the browser, but all without success. Does anyone of you have an idea how I can solve the problem? ### Outcome #### What did you expect? #### What happened instead? ### Your phone model iPhone12, iPhone14 ProMax, iPhone6s, iPad Air ### Operating system version iOS16.2 ### Application version _No response_ ### Homeserver yatrix.org ### Will you send logs? No
1.0
Notification on iOS (wrong numbers of notifiaction in the red bubble) - ### Steps to reproduce I have the problem that every message that the Element Messenger receives shows a three in the notification bubble instead of a one. I have now tried all possible notification options both in the iOS app and directly in the browser, but all without success. Does anyone of you have an idea how I can solve the problem? ### Outcome #### What did you expect? #### What happened instead? ### Your phone model iPhone12, iPhone14 ProMax, iPhone6s, iPad Air ### Operating system version iOS16.2 ### Application version _No response_ ### Homeserver yatrix.org ### Will you send logs? No
defect
notification on ios wrong numbers of notifiaction in the red bubble steps to reproduce i have the problem that every message that the element messenger receives shows a three in the notification bubble instead of a one i have now tried all possible notification options both in the ios app and directly in the browser but all without success does anyone of you have an idea how i can solve the problem outcome what did you expect what happened instead your phone model promax ipad air operating system version application version no response homeserver yatrix org will you send logs no
1
117,541
4,718,203,183
IssuesEvent
2016-10-17 00:16:56
nathan-russell/hashmap
https://api.github.com/repos/nathan-russell/hashmap
closed
Set attributes for uncached calls to data
enhancement priority-low
Most likely needs to happen [before coercion](https://github.com/nathan-russell/hashmap/blob/master/inst/include/hashmap/HashTemplate.hpp#L364-L374). ```R library(hashmap) x <- Sys.Date() + 1:5 y <- Sys.Date() + 1:5 h <- hashmap(x, y) h$data() # 16911 16910 16912 16909 16908 # "2016-04-20" "2016-04-19" "2016-04-21" "2016-04-18" "2016-04-17" h$cache_keys() h$cache_values() h$data() # 2016-04-20 2016-04-19 2016-04-21 2016-04-18 2016-04-17 # "2016-04-20" "2016-04-19" "2016-04-21" "2016-04-18" "2016-04-17" ## ## x <- Sys.time() + 3600 * (1:5) y <- Sys.time() + 3600 * (1:5) h <- hashmap(x, y) h$data() # 1460841642.22724 1460834442.22724 1460838042.22724 # "2016-04-16 21:20:52 UTC" "2016-04-16 19:20:52 UTC" "2016-04-16 20:20:52 UTC" # 1460830842.22724 1460827242.22724 # "2016-04-16 18:20:52 UTC" "2016-04-16 17:20:52 UTC" h$cache_keys() h$cache_values() h$data() # 2016-04-16 21:20:42 2016-04-16 19:20:42 2016-04-16 20:20:42 # "2016-04-16 21:20:52 UTC" "2016-04-16 19:20:52 UTC" "2016-04-16 20:20:52 UTC" # 2016-04-16 18:20:42 2016-04-16 17:20:42 # "2016-04-16 18:20:52 UTC" "2016-04-16 17:20:52 UTC" ```
1.0
Set attributes for uncached calls to data - Most likely needs to happen [before coercion](https://github.com/nathan-russell/hashmap/blob/master/inst/include/hashmap/HashTemplate.hpp#L364-L374). ```R library(hashmap) x <- Sys.Date() + 1:5 y <- Sys.Date() + 1:5 h <- hashmap(x, y) h$data() # 16911 16910 16912 16909 16908 # "2016-04-20" "2016-04-19" "2016-04-21" "2016-04-18" "2016-04-17" h$cache_keys() h$cache_values() h$data() # 2016-04-20 2016-04-19 2016-04-21 2016-04-18 2016-04-17 # "2016-04-20" "2016-04-19" "2016-04-21" "2016-04-18" "2016-04-17" ## ## x <- Sys.time() + 3600 * (1:5) y <- Sys.time() + 3600 * (1:5) h <- hashmap(x, y) h$data() # 1460841642.22724 1460834442.22724 1460838042.22724 # "2016-04-16 21:20:52 UTC" "2016-04-16 19:20:52 UTC" "2016-04-16 20:20:52 UTC" # 1460830842.22724 1460827242.22724 # "2016-04-16 18:20:52 UTC" "2016-04-16 17:20:52 UTC" h$cache_keys() h$cache_values() h$data() # 2016-04-16 21:20:42 2016-04-16 19:20:42 2016-04-16 20:20:42 # "2016-04-16 21:20:52 UTC" "2016-04-16 19:20:52 UTC" "2016-04-16 20:20:52 UTC" # 2016-04-16 18:20:42 2016-04-16 17:20:42 # "2016-04-16 18:20:52 UTC" "2016-04-16 17:20:52 UTC" ```
non_defect
set attributes for uncached calls to data most likely needs to happen r library hashmap x sys date y sys date h hashmap x y h data h cache keys h cache values h data x sys time y sys time h hashmap x y h data utc utc utc utc utc h cache keys h cache values h data utc utc utc utc utc
0
43,797
11,849,853,526
IssuesEvent
2020-03-24 15:49:02
mestrade/go-hello
https://api.github.com/repos/mestrade/go-hello
opened
Unauthorized API Usage
security/defectDojo
*Unauthorized API Usage* *Severity:* Info *Cve:* *Product/Engagement:* test / Ad Hoc Engagement *Systems*: *Description*: Long description *Mitigation*: mitigation info *Impact*: impact is here *References*:No references given
1.0
Unauthorized API Usage - *Unauthorized API Usage* *Severity:* Info *Cve:* *Product/Engagement:* test / Ad Hoc Engagement *Systems*: *Description*: Long description *Mitigation*: mitigation info *Impact*: impact is here *References*:No references given
defect
unauthorized api usage unauthorized api usage severity info cve product engagement test ad hoc engagement systems description long description mitigation mitigation info impact impact is here references no references given
1
52,814
13,225,094,697
IssuesEvent
2020-08-17 20:28:43
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
GLShovel does not correctly show pulses/launches/i3particle if the viewing timewindow is not in range (Trac #390)
Migrated from Trac defect glshovel
GLShovel has problems with long hitseries on a any DOM, when the viewing time excludes the first hit. If for example a hitseries has two hits, 1 at 10 microseconds and one at 100 microseconds, and we put the t_min to 50 and t_max to 200, no hit will be rendered for that DOM. The same is true for domlaunches and I3Particles. For example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event, you have to back in time often quite far to be able to see the track. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/390">https://code.icecube.wisc.edu/projects/icecube/ticket/390</a>, reported by gluesenkampand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2013-11-21T22:34:18", "_ts": "1385073258000000", "description": "GLShovel has problems with long hitseries on a any DOM, when the viewing time excludes the first hit.\n\nIf for example a hitseries has two hits, 1 at 10 microseconds and one\nat 100 microseconds, and we put the t_min to 50 and t_max to 200,\nno hit will be rendered for that DOM.\n\n\nThe same is true for domlaunches and I3Particles. For example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event, you have to back in time often quite far to be able to see the track.\n\n", "reporter": "gluesenkamp", "cc": "", "resolution": "fixed", "time": "2012-04-26T15:11:55", "component": "glshovel", "summary": "GLShovel does not correctly show pulses/launches/i3particle if the viewing timewindow is not in range", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
GLShovel does not correctly show pulses/launches/i3particle if the viewing timewindow is not in range (Trac #390) - GLShovel has problems with long hitseries on a any DOM, when the viewing time excludes the first hit. If for example a hitseries has two hits, 1 at 10 microseconds and one at 100 microseconds, and we put the t_min to 50 and t_max to 200, no hit will be rendered for that DOM. The same is true for domlaunches and I3Particles. For example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event, you have to back in time often quite far to be able to see the track. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/390">https://code.icecube.wisc.edu/projects/icecube/ticket/390</a>, reported by gluesenkampand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2013-11-21T22:34:18", "_ts": "1385073258000000", "description": "GLShovel has problems with long hitseries on a any DOM, when the viewing time excludes the first hit.\n\nIf for example a hitseries has two hits, 1 at 10 microseconds and one\nat 100 microseconds, and we put the t_min to 50 and t_max to 200,\nno hit will be rendered for that DOM.\n\n\nThe same is true for domlaunches and I3Particles. For example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event, you have to back in time often quite far to be able to see the track.\n\n", "reporter": "gluesenkamp", "cc": "", "resolution": "fixed", "time": "2012-04-26T15:11:55", "component": "glshovel", "summary": "GLShovel does not correctly show pulses/launches/i3particle if the viewing timewindow is not in range", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
glshovel does not correctly show pulses launches if the viewing timewindow is not in range trac glshovel has problems with long hitseries on a any dom when the viewing time excludes the first hit if for example a hitseries has two hits at microseconds and one at microseconds and we put the t min to and t max to no hit will be rendered for that dom the same is true for domlaunches and for example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event you have to back in time often quite far to be able to see the track migrated from json status closed changetime ts description glshovel has problems with long hitseries on a any dom when the viewing time excludes the first hit n nif for example a hitseries has two hits at microseconds and one nat microseconds and we put the t min to and t max to nno hit will be rendered for that dom n n nthe same is true for domlaunches and for example for neutrino datasets that come through earth where the neutrino is much earlier than the actual triggered event you have to back in time often quite far to be able to see the track n n reporter gluesenkamp cc resolution fixed time component glshovel summary glshovel does not correctly show pulses launches if the viewing timewindow is not in range priority normal keywords milestone owner olivas type defect
1
66,564
20,328,505,256
IssuesEvent
2022-02-18 08:28:03
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
ZFS doesn't respect Linux kernel CPU isolation mechanisms
Type: Defect Status: Stale
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | ArchLinux Distribution Version | Rolling Linux Kernel | 4.19.48 Architecture | x86_64 ZFS Version | 0.8.0 SPL Version | 0.8.0 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing `module/spl/spl-taskq.c` contains this code: ```c tqt->tqt_thread = spl_kthread_create(taskq_thread, tqt, "%s", tq->tq_name); if (tqt->tqt_thread == NULL) { kmem_free(tqt, sizeof (taskq_thread_t)); return (NULL); } if (spl_taskq_thread_bind) { last_used_cpu = (last_used_cpu + 1) % num_online_cpus(); kthread_bind(tqt->tqt_thread, last_used_cpu); } ``` Thus, kthreads spawn either with the default cpumask or, if `spl_taskq_thread_bind=1` is set on module import, are bound to CPUs without regard for their availability to the scheduler. This can be a substantial source of latency, which is not acceptable on many systems that use the `isolcpus` boot parameter to isolate designated "real-time" cores. While `spl_taskq_thread_bind=1` prevents latency from thread migration on/off RT CPUs, it can make things substantially worse by locking the threads to arbitrary cores in a way that can't be changed with `taskset`, leaving the RT CPUs permanently saddled with the kthread for its full lifecycle. Ideally, the modular CPU selection would be replaced with something that uses the kernel's housekeeping API in `include/linux/sched/isolation.h` to get the cpumask of non-isolated CPUs and use `kthread_create_on_cpu` in `spl_kthread_create` and/or ~~`kthread_bind_mask`~~ to schedule and bind threads across non-RT cores only. Note, however, this is an incomplete solution because the kernel's interface to get an `isolcpus` cpumask has changed several times across the versions supported by ZFS. Various hacks can be done to try to prevent unbound kthreads from using isolated cores, and threads not bound with `spl_taskq_thread_bind` can be moved, but these solutions are iffy and incomplete at best. It would be great if ZFS respected `isolcpus` from the start. ### Describe how to reproduce the problem Boot with `isolcpus`, capture a trace of the RT CPUs with `perf sched record` or other tracing mechanisms, observe ZFS-spawned kthreads coming on and off isolated cores. This is the primary remaining source of latency on my local system. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
ZFS doesn't respect Linux kernel CPU isolation mechanisms - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | ArchLinux Distribution Version | Rolling Linux Kernel | 4.19.48 Architecture | x86_64 ZFS Version | 0.8.0 SPL Version | 0.8.0 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing `module/spl/spl-taskq.c` contains this code: ```c tqt->tqt_thread = spl_kthread_create(taskq_thread, tqt, "%s", tq->tq_name); if (tqt->tqt_thread == NULL) { kmem_free(tqt, sizeof (taskq_thread_t)); return (NULL); } if (spl_taskq_thread_bind) { last_used_cpu = (last_used_cpu + 1) % num_online_cpus(); kthread_bind(tqt->tqt_thread, last_used_cpu); } ``` Thus, kthreads spawn either with the default cpumask or, if `spl_taskq_thread_bind=1` is set on module import, are bound to CPUs without regard for their availability to the scheduler. This can be a substantial source of latency, which is not acceptable on many systems that use the `isolcpus` boot parameter to isolate designated "real-time" cores. While `spl_taskq_thread_bind=1` prevents latency from thread migration on/off RT CPUs, it can make things substantially worse by locking the threads to arbitrary cores in a way that can't be changed with `taskset`, leaving the RT CPUs permanently saddled with the kthread for its full lifecycle. Ideally, the modular CPU selection would be replaced with something that uses the kernel's housekeeping API in `include/linux/sched/isolation.h` to get the cpumask of non-isolated CPUs and use `kthread_create_on_cpu` in `spl_kthread_create` and/or ~~`kthread_bind_mask`~~ to schedule and bind threads across non-RT cores only. Note, however, this is an incomplete solution because the kernel's interface to get an `isolcpus` cpumask has changed several times across the versions supported by ZFS. Various hacks can be done to try to prevent unbound kthreads from using isolated cores, and threads not bound with `spl_taskq_thread_bind` can be moved, but these solutions are iffy and incomplete at best. It would be great if ZFS respected `isolcpus` from the start. ### Describe how to reproduce the problem Boot with `isolcpus`, capture a trace of the RT CPUs with `perf sched record` or other tracing mechanisms, observe ZFS-spawned kthreads coming on and off isolated cores. This is the primary remaining source of latency on my local system. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
zfs doesn t respect linux kernel cpu isolation mechanisms thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information type version name distribution name archlinux distribution version rolling linux kernel architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing module spl spl taskq c contains this code c tqt tqt thread spl kthread create taskq thread tqt s tq tq name if tqt tqt thread null kmem free tqt sizeof taskq thread t return null if spl taskq thread bind last used cpu last used cpu num online cpus kthread bind tqt tqt thread last used cpu thus kthreads spawn either with the default cpumask or if spl taskq thread bind is set on module import are bound to cpus without regard for their availability to the scheduler this can be a substantial source of latency which is not acceptable on many systems that use the isolcpus boot parameter to isolate designated real time cores while spl taskq thread bind prevents latency from thread migration on off rt cpus it can make things substantially worse by locking the threads to arbitrary cores in a way that can t be changed with taskset leaving the rt cpus permanently saddled with the kthread for its full lifecycle ideally the modular cpu selection would be replaced with something that uses the kernel s housekeeping api in include linux sched isolation h to get the cpumask of non isolated cpus and use kthread create on cpu in spl kthread create and or kthread bind mask to schedule and bind threads across non rt cores only note however this is an incomplete solution because the kernel s interface to get an isolcpus cpumask has changed several times across the versions supported by zfs various hacks can be done to try to prevent unbound kthreads from using isolated cores and threads not bound with spl taskq thread bind can be moved but these solutions are iffy and incomplete at best it would be great if zfs respected isolcpus from the start describe how to reproduce the problem boot with isolcpus capture a trace of the rt cpus with perf sched record or other tracing mechanisms observe zfs spawned kthreads coming on and off isolated cores this is the primary remaining source of latency on my local system include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
12,981
2,732,346,280
IssuesEvent
2015-04-17 04:48:04
rasmus/fast-member
https://api.github.com/repos/rasmus/fast-member
closed
Access to static properties and fields
auto-migrated Priority-Medium Type-Defect
``` How can we make ti working with static properties? Here is a sample to product problem: var accessor = TypeAccessor.Create(typeof (DateTime)); var now = accessor[null, "Now"]; Regards, ``` Original issue reported on code.google.com by `shama...@gmail.com` on 20 Nov 2014 at 10:50
1.0
Access to static properties and fields - ``` How can we make ti working with static properties? Here is a sample to product problem: var accessor = TypeAccessor.Create(typeof (DateTime)); var now = accessor[null, "Now"]; Regards, ``` Original issue reported on code.google.com by `shama...@gmail.com` on 20 Nov 2014 at 10:50
defect
access to static properties and fields how can we make ti working with static properties here is a sample to product problem var accessor typeaccessor create typeof datetime var now accessor regards original issue reported on code google com by shama gmail com on nov at
1
347,807
31,276,593,007
IssuesEvent
2023-08-22 06:47:43
farhanramzy/upptime
https://api.github.com/repos/farhanramzy/upptime
opened
🛑 Test Broken real is down
status test-broken-real
In [`065cf01`](https://github.com/farhanramzy/upptime/commit/065cf01862c8e8df9a42eb6e3cf8b1e36c58e662 ), Test Broken real (https://thissitedoesnotexist.koj.co) was **down**: - HTTP code: 0 - Response time: 0 ms
1.0
🛑 Test Broken real is down - In [`065cf01`](https://github.com/farhanramzy/upptime/commit/065cf01862c8e8df9a42eb6e3cf8b1e36c58e662 ), Test Broken real (https://thissitedoesnotexist.koj.co) was **down**: - HTTP code: 0 - Response time: 0 ms
non_defect
🛑 test broken real is down in test broken real was down http code response time ms
0
41,202
10,332,206,388
IssuesEvent
2019-09-02 22:12:46
MarcusWolschon/osmeditor4android
https://api.github.com/repos/MarcusWolschon/osmeditor4android
closed
Rotate Tool Buggy
Defect
Description: --- (Pixel 3A XL) I was using the Rotate Tool and while I was adjusting the Way to turn it 180 degrees the Way would disappear. So, I kept undoing back to before I chose the tool and eventually I got the result I was looking for, but now the Way didn't have any Nodes anymore. I ended up giving up using the Rotate Tool and drew the shape I wanted to create. Later when I tried to upload the changes I was getting an API error over and over again that kept me from uploading. I went through the log (before you upload) and found that there was some hidden shapes that disappeared when I was trying to rotate. So I removed the hidden shape and now I was able to upload the changes, but now the map has disappeared and I'm getting a black screen (I can still interact with the selections at the bottom of the screen). I'm probably going to uninstall and reinstall the app to fix this problem, but wanted you to know. Device info: --- <table> <tr><td>App version</td><td>13.1.6.0</td></tr> <tr><td>App version code</td><td>1209</td></tr> <tr><td>Android build version</td><td>5674421</td></tr> <tr><td>Android release version</td><td>9</td></tr> <tr><td>Android SDK version</td><td>28</td></tr> <tr><td>Android build ID</td><td>PQ3B.190801.002</td></tr> <tr><td>Device brand</td><td>google</td></tr> <tr><td>Device manufacturer</td><td>Google</td></tr> <tr><td>Device name</td><td>bonito</td></tr> <tr><td>Device model</td><td>Pixel 3a XL</td></tr> <tr><td>Device product name</td><td>bonito</td></tr> <tr><td>Device hardware name</td><td>bonito</td></tr> <tr><td>ABIs</td><td>[arm64-v8a, armeabi-v7a, armeabi]</td></tr> <tr><td>ABIs (32bit)</td><td>[armeabi-v7a, armeabi]</td></tr> <tr><td>ABIs (64bit)</td><td>[arm64-v8a]</td></tr> </table> Extra info: --- <table> <tr><td>OSM display name</td><td>MapRogers</td></tr> </table>
1.0
Rotate Tool Buggy - Description: --- (Pixel 3A XL) I was using the Rotate Tool and while I was adjusting the Way to turn it 180 degrees the Way would disappear. So, I kept undoing back to before I chose the tool and eventually I got the result I was looking for, but now the Way didn't have any Nodes anymore. I ended up giving up using the Rotate Tool and drew the shape I wanted to create. Later when I tried to upload the changes I was getting an API error over and over again that kept me from uploading. I went through the log (before you upload) and found that there was some hidden shapes that disappeared when I was trying to rotate. So I removed the hidden shape and now I was able to upload the changes, but now the map has disappeared and I'm getting a black screen (I can still interact with the selections at the bottom of the screen). I'm probably going to uninstall and reinstall the app to fix this problem, but wanted you to know. Device info: --- <table> <tr><td>App version</td><td>13.1.6.0</td></tr> <tr><td>App version code</td><td>1209</td></tr> <tr><td>Android build version</td><td>5674421</td></tr> <tr><td>Android release version</td><td>9</td></tr> <tr><td>Android SDK version</td><td>28</td></tr> <tr><td>Android build ID</td><td>PQ3B.190801.002</td></tr> <tr><td>Device brand</td><td>google</td></tr> <tr><td>Device manufacturer</td><td>Google</td></tr> <tr><td>Device name</td><td>bonito</td></tr> <tr><td>Device model</td><td>Pixel 3a XL</td></tr> <tr><td>Device product name</td><td>bonito</td></tr> <tr><td>Device hardware name</td><td>bonito</td></tr> <tr><td>ABIs</td><td>[arm64-v8a, armeabi-v7a, armeabi]</td></tr> <tr><td>ABIs (32bit)</td><td>[armeabi-v7a, armeabi]</td></tr> <tr><td>ABIs (64bit)</td><td>[arm64-v8a]</td></tr> </table> Extra info: --- <table> <tr><td>OSM display name</td><td>MapRogers</td></tr> </table>
defect
rotate tool buggy description pixel xl i was using the rotate tool and while i was adjusting the way to turn it degrees the way would disappear so i kept undoing back to before i chose the tool and eventually i got the result i was looking for but now the way didn t have any nodes anymore i ended up giving up using the rotate tool and drew the shape i wanted to create later when i tried to upload the changes i was getting an api error over and over again that kept me from uploading i went through the log before you upload and found that there was some hidden shapes that disappeared when i was trying to rotate so i removed the hidden shape and now i was able to upload the changes but now the map has disappeared and i m getting a black screen i can still interact with the selections at the bottom of the screen i m probably going to uninstall and reinstall the app to fix this problem but wanted you to know device info app version app version code android build version android release version android sdk version android build id device brand google device manufacturer google device name bonito device model pixel xl device product name bonito device hardware name bonito abis abis abis extra info osm display name maprogers
1
132,943
10,774,066,575
IssuesEvent
2019-11-03 01:42:18
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
BUG: replace int64 dtype with bool coerces to int64
Needs Tests good first issue
``` # NG s = pd.Series([1, 2, 3]) s.replace(2, True) #0 1 #1 1 #2 3 # dtype: int64 ``` If target has `object` dtype, it works. ``` s = pd.Series(['x', 2, 3]) s.replace(2, True) #0 x #1 True #2 3 # dtype: object ```
1.0
BUG: replace int64 dtype with bool coerces to int64 - ``` # NG s = pd.Series([1, 2, 3]) s.replace(2, True) #0 1 #1 1 #2 3 # dtype: int64 ``` If target has `object` dtype, it works. ``` s = pd.Series(['x', 2, 3]) s.replace(2, True) #0 x #1 True #2 3 # dtype: object ```
non_defect
bug replace dtype with bool coerces to ng s pd series s replace true dtype if target has object dtype it works s pd series s replace true x true dtype object
0
32,708
6,898,727,707
IssuesEvent
2017-11-24 10:40:23
veekoon/droidsheep-2.0
https://api.github.com/repos/veekoon/droidsheep-2.0
closed
Android L not supported
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Install the application on android L 2. Start spoofing 3. Check the SuperSu logs fr error. Android L supports PIE applications only. What is the expected output? What do you see instead? It would be great if DroidSheep was modifid to match PIE What version of the product are you using? On what operating system? Android 5 Please provide any additional information below. ``` Original issue reported on code.google.com by `jjjh...@gmail.com` on 6 Dec 2014 at 8:40
1.0
Android L not supported - ``` What steps will reproduce the problem? 1. Install the application on android L 2. Start spoofing 3. Check the SuperSu logs fr error. Android L supports PIE applications only. What is the expected output? What do you see instead? It would be great if DroidSheep was modifid to match PIE What version of the product are you using? On what operating system? Android 5 Please provide any additional information below. ``` Original issue reported on code.google.com by `jjjh...@gmail.com` on 6 Dec 2014 at 8:40
defect
android l not supported what steps will reproduce the problem install the application on android l start spoofing check the supersu logs fr error android l supports pie applications only what is the expected output what do you see instead it would be great if droidsheep was modifid to match pie what version of the product are you using on what operating system android please provide any additional information below original issue reported on code google com by jjjh gmail com on dec at
1
42,802
11,272,452,519
IssuesEvent
2020-01-14 14:54:40
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Possible resource leaks related to ClusterViewListenerService.clusterListeningEndpoints
Team: Client Team: Core Type: Defect
While investigating a Jet OOME, we've realized there can be issues related to `ClusterViewListenerService`. First of all, there were cca 27000 `TcpIpConnection` instances, but just about 20 of them were alive. The size of `clusterListeningEndpoints` map was about 25000 entries. (@olukas can provide the original heapdump). We didn't find the reason for this behavior, but we realized there can be a resource leak in `com.hazelcast.client.impl.ClientEndpointImpl.destroy()` E.g. when a `logout` throws a `LoginException` or another problem jumps in before removing endpoint from the map. https://github.com/hazelcast/hazelcast/blob/fdc6a205b2b6f50e01d21a263c373a0894ad2e67/hazelcast/src/main/java/com/hazelcast/client/impl/ClientEndpointImpl.java#L214-L231
1.0
Possible resource leaks related to ClusterViewListenerService.clusterListeningEndpoints - While investigating a Jet OOME, we've realized there can be issues related to `ClusterViewListenerService`. First of all, there were cca 27000 `TcpIpConnection` instances, but just about 20 of them were alive. The size of `clusterListeningEndpoints` map was about 25000 entries. (@olukas can provide the original heapdump). We didn't find the reason for this behavior, but we realized there can be a resource leak in `com.hazelcast.client.impl.ClientEndpointImpl.destroy()` E.g. when a `logout` throws a `LoginException` or another problem jumps in before removing endpoint from the map. https://github.com/hazelcast/hazelcast/blob/fdc6a205b2b6f50e01d21a263c373a0894ad2e67/hazelcast/src/main/java/com/hazelcast/client/impl/ClientEndpointImpl.java#L214-L231
defect
possible resource leaks related to clusterviewlistenerservice clusterlisteningendpoints while investigating a jet oome we ve realized there can be issues related to clusterviewlistenerservice first of all there were cca tcpipconnection instances but just about of them were alive the size of clusterlisteningendpoints map was about entries olukas can provide the original heapdump we didn t find the reason for this behavior but we realized there can be a resource leak in com hazelcast client impl clientendpointimpl destroy e g when a logout throws a loginexception or another problem jumps in before removing endpoint from the map
1
20,928
3,436,685,711
IssuesEvent
2015-12-12 16:09:38
numpy/numpy
https://api.github.com/repos/numpy/numpy
closed
Windows testing errors for 1.10.2
5 - Testing component: numpy.ma component: numpy.random Defect
Christoph reports on 1.10.x On Python 2.x, 32 and 64 bit: ``` ================================================== FAIL: test_mvoid_multidim_print (test_core.TestMaskedArray) ---------------------------------------------------------------------- Traceback (most recent call last): File "X:\Python27-x64\lib\site-packages\numpy\ma\tests\test_core.py", line 716, in test_mvoid_multidim_print assert str(t_ma[0]) == "([1, --, 3],)" AssertionError ``` On Python 2.x, 32 bit: ``` =================================================== FAIL: test_umath.test_rint_big_int ---------------------------------------------------------------------- Traceback (most recent call last): File "X:\Python27\lib\site-packages\nose\case.py", line 197, in runTest self.test(*self.arg) File "X:\Python27\lib\site-packages\numpy\core\tests\test_umath.py", line 1939, in test_rint_big_int assert_equal(val, np.rint(val)) File "X:\Python27\lib\site-packages\numpy\testing\utils.py", line 355, in assert_equal raise AssertionError(msg) AssertionError: Items are not equal: ACTUAL: 4607998452777363968L DESIRED: 4.6079984527773635e+18 ``` The second looks like a promotion of `long` to `float`, not sure about the first. Can someone running windows track these down? I don't think they are serious, but it would be good to know where they come from. Note that in the first test `assert_` is not used.
1.0
Windows testing errors for 1.10.2 - Christoph reports on 1.10.x On Python 2.x, 32 and 64 bit: ``` ================================================== FAIL: test_mvoid_multidim_print (test_core.TestMaskedArray) ---------------------------------------------------------------------- Traceback (most recent call last): File "X:\Python27-x64\lib\site-packages\numpy\ma\tests\test_core.py", line 716, in test_mvoid_multidim_print assert str(t_ma[0]) == "([1, --, 3],)" AssertionError ``` On Python 2.x, 32 bit: ``` =================================================== FAIL: test_umath.test_rint_big_int ---------------------------------------------------------------------- Traceback (most recent call last): File "X:\Python27\lib\site-packages\nose\case.py", line 197, in runTest self.test(*self.arg) File "X:\Python27\lib\site-packages\numpy\core\tests\test_umath.py", line 1939, in test_rint_big_int assert_equal(val, np.rint(val)) File "X:\Python27\lib\site-packages\numpy\testing\utils.py", line 355, in assert_equal raise AssertionError(msg) AssertionError: Items are not equal: ACTUAL: 4607998452777363968L DESIRED: 4.6079984527773635e+18 ``` The second looks like a promotion of `long` to `float`, not sure about the first. Can someone running windows track these down? I don't think they are serious, but it would be good to know where they come from. Note that in the first test `assert_` is not used.
defect
windows testing errors for christoph reports on x on python x and bit fail test mvoid multidim print test core testmaskedarray traceback most recent call last file x lib site packages numpy ma tests test core py line in test mvoid multidim print assert str t ma assertionerror on python x bit fail test umath test rint big int traceback most recent call last file x lib site packages nose case py line in runtest self test self arg file x lib site packages numpy core tests test umath py line in test rint big int assert equal val np rint val file x lib site packages numpy testing utils py line in assert equal raise assertionerror msg assertionerror items are not equal actual desired the second looks like a promotion of long to float not sure about the first can someone running windows track these down i don t think they are serious but it would be good to know where they come from note that in the first test assert is not used
1
321,965
9,810,937,290
IssuesEvent
2019-06-12 21:49:54
DlfinBroom/ChatBot
https://api.github.com/repos/DlfinBroom/ChatBot
closed
Finish About Page
Finish Priority: Low View/Page
Needs to include a few paragraphs about this project and a link to the contact page
1.0
Finish About Page - Needs to include a few paragraphs about this project and a link to the contact page
non_defect
finish about page needs to include a few paragraphs about this project and a link to the contact page
0
36,488
7,963,427,548
IssuesEvent
2018-07-13 17:33:02
CocoaPods/CocoaPods
https://api.github.com/repos/CocoaPods/CocoaPods
closed
Cannot run unit tests in new Swift pod
help wanted s7:workaround available t2:defect
* [X] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report ## What did you do? - `pod lib create Rules` - Answered all the questions with the default answers except say No to view based testing. - Added my Swift files to the project's Classes folder. - `pod install` - Open the workspace. - Cmd-U (Test). ## What did you expect to happen? I expected tests to run. ## What happened instead? Build succeeds, but with this warning: ``` ld: warning: directory not found for option '-F/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator11.4.sdk/Developer/Library/Frameworks' ``` And this console output: ``` 2018-06-28 16:55:48.943967-0700 xctest[39973:4152757] The bundle “Rules_Tests” couldn’t be loaded because it is damaged or missing necessary resources. Try reinstalling the bundle. 2018-06-28 16:55:48.944105-0700 xctest[39973:4152757] (dlopen_preflight(/Users/jim/Library/Developer/Xcode/DerivedData/Rules-cymfcyukficsvsghhxyuhzarvneb/Build/Products/Debug-iphonesimulator/Rules_Tests.xctest/Rules_Tests): Library not loaded: @rpath/Rules.framework/Rules Referenced from: /Users/jim/Library/Developer/Xcode/DerivedData/Rules-cymfcyukficsvsghhxyuhzarvneb/Build/Products/Debug-iphonesimulator/Rules_Tests.xctest/Rules_Tests Reason: image not found) Program ended with exit code: 82 ``` ## CocoaPods Environment ### Stack ``` CocoaPods : 1.5.3 Ruby : ruby 2.3.1p112 (2016-04-26 revision 54768) [x86_64-darwin15] RubyGems : 2.5.1 Host : Mac OS X 10.13.4 (17E199) Xcode : 9.4 (9F1027a) Git : git version 2.10.1 Ruby lib dir : /Users/jim/.rbenv/versions/2.3.1/lib Repositories : master - https://github.com/CocoaPods/Specs.git @ 8c2aa947e5c36c90ca3bb7fa510fc9b0474a9b9d ``` ### Installation Source ``` Executable Path: /Users/jim/.gem/bin/pod ``` ### Plugins ``` claide-plugins : 0.9.2 cocoapods-deintegrate : 1.0.2 cocoapods-keys : 1.7.0 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.0.0 cocoapods-trunk : 1.3.0 cocoapods-try : 1.1.0 cocoapods-update-if-you-dare : 0.1.1 slather : 2.3.0 ``` ### Podfile ```ruby use_frameworks! target 'Rules_Example' do pod 'Rules', :path => '../' target 'Rules_Tests' do inherit! :search_paths pod 'Quick', '~> 1.2.0' pod 'Nimble', '~> 7.0.2' end end ```
1.0
Cannot run unit tests in new Swift pod - * [X] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report ## What did you do? - `pod lib create Rules` - Answered all the questions with the default answers except say No to view based testing. - Added my Swift files to the project's Classes folder. - `pod install` - Open the workspace. - Cmd-U (Test). ## What did you expect to happen? I expected tests to run. ## What happened instead? Build succeeds, but with this warning: ``` ld: warning: directory not found for option '-F/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator11.4.sdk/Developer/Library/Frameworks' ``` And this console output: ``` 2018-06-28 16:55:48.943967-0700 xctest[39973:4152757] The bundle “Rules_Tests” couldn’t be loaded because it is damaged or missing necessary resources. Try reinstalling the bundle. 2018-06-28 16:55:48.944105-0700 xctest[39973:4152757] (dlopen_preflight(/Users/jim/Library/Developer/Xcode/DerivedData/Rules-cymfcyukficsvsghhxyuhzarvneb/Build/Products/Debug-iphonesimulator/Rules_Tests.xctest/Rules_Tests): Library not loaded: @rpath/Rules.framework/Rules Referenced from: /Users/jim/Library/Developer/Xcode/DerivedData/Rules-cymfcyukficsvsghhxyuhzarvneb/Build/Products/Debug-iphonesimulator/Rules_Tests.xctest/Rules_Tests Reason: image not found) Program ended with exit code: 82 ``` ## CocoaPods Environment ### Stack ``` CocoaPods : 1.5.3 Ruby : ruby 2.3.1p112 (2016-04-26 revision 54768) [x86_64-darwin15] RubyGems : 2.5.1 Host : Mac OS X 10.13.4 (17E199) Xcode : 9.4 (9F1027a) Git : git version 2.10.1 Ruby lib dir : /Users/jim/.rbenv/versions/2.3.1/lib Repositories : master - https://github.com/CocoaPods/Specs.git @ 8c2aa947e5c36c90ca3bb7fa510fc9b0474a9b9d ``` ### Installation Source ``` Executable Path: /Users/jim/.gem/bin/pod ``` ### Plugins ``` claide-plugins : 0.9.2 cocoapods-deintegrate : 1.0.2 cocoapods-keys : 1.7.0 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.0.0 cocoapods-trunk : 1.3.0 cocoapods-try : 1.1.0 cocoapods-update-if-you-dare : 0.1.1 slather : 2.3.0 ``` ### Podfile ```ruby use_frameworks! target 'Rules_Example' do pod 'Rules', :path => '../' target 'Rules_Tests' do inherit! :search_paths pod 'Quick', '~> 1.2.0' pod 'Nimble', '~> 7.0.2' end end ```
defect
cannot run unit tests in new swift pod i ve read and understood the report what did you do pod lib create rules answered all the questions with the default answers except say no to view based testing added my swift files to the project s classes folder pod install open the workspace cmd u test what did you expect to happen i expected tests to run what happened instead build succeeds but with this warning ld warning directory not found for option f applications xcode app contents developer platforms iphonesimulator platform developer sdks sdk developer library frameworks and this console output xctest the bundle “rules tests” couldn’t be loaded because it is damaged or missing necessary resources try reinstalling the bundle xctest dlopen preflight users jim library developer xcode deriveddata rules cymfcyukficsvsghhxyuhzarvneb build products debug iphonesimulator rules tests xctest rules tests library not loaded rpath rules framework rules referenced from users jim library developer xcode deriveddata rules cymfcyukficsvsghhxyuhzarvneb build products debug iphonesimulator rules tests xctest rules tests reason image not found program ended with exit code cocoapods environment stack cocoapods ruby ruby revision rubygems host mac os x xcode git git version ruby lib dir users jim rbenv versions lib repositories master installation source executable path users jim gem bin pod plugins claide plugins cocoapods deintegrate cocoapods keys cocoapods plugins cocoapods search cocoapods stats cocoapods trunk cocoapods try cocoapods update if you dare slather podfile ruby use frameworks target rules example do pod rules path target rules tests do inherit search paths pod quick pod nimble end end
1
398,493
11,741,573,497
IssuesEvent
2020-03-11 22:05:15
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
Change Events in a locality to verification status = "accepted"
Function-Locality/Event/Georeferencing Priority-High
I need to change all specimen events associated with this locality to "accepted" https://arctos.database.museum/SpecimenResults.cfm?locality_id=11025908 There are more than 1000 events, so I cannot use the tool in specimen results. @dustymc can you do this one for me? I'd like to get these changed before I add the public events.
1.0
Change Events in a locality to verification status = "accepted" - I need to change all specimen events associated with this locality to "accepted" https://arctos.database.museum/SpecimenResults.cfm?locality_id=11025908 There are more than 1000 events, so I cannot use the tool in specimen results. @dustymc can you do this one for me? I'd like to get these changed before I add the public events.
non_defect
change events in a locality to verification status accepted i need to change all specimen events associated with this locality to accepted there are more than events so i cannot use the tool in specimen results dustymc can you do this one for me i d like to get these changed before i add the public events
0
19,082
3,136,120,080
IssuesEvent
2015-09-10 18:21:42
GoldenSoftwareLtd/GedeminSalary
https://api.github.com/repos/GoldenSoftwareLtd/GedeminSalary
closed
Сумма изменения зарплаты
Priority-High Type-Defect
Неверно рассчитывается % изменение зарплаты относительно предыдущего месяца. Пример есть.
1.0
Сумма изменения зарплаты - Неверно рассчитывается % изменение зарплаты относительно предыдущего месяца. Пример есть.
defect
сумма изменения зарплаты неверно рассчитывается изменение зарплаты относительно предыдущего месяца пример есть
1
109,626
23,800,965,496
IssuesEvent
2022-09-03 09:25:01
Toma400/The_Isle_of_Ansur
https://api.github.com/repos/Toma400/The_Isle_of_Ansur
opened
Better "prioritised" system for panoramas and menu sounds & Mod Loading Order
feature suggestion code improvement
Having possibility to prioritise panoramas and sounds is cool, as they will overwrite vanilla, but... They will not overwrite other mods. To explain it a bit in detail: if more than one mod use "PR%_" system, they will both shuffle through their data, so more than one mod can be prioritised. It's okay if you just want to overwrite vanilla, but not if you want to rule over all possible mods. Solution for this can be kinda simple, kinda not, mechanic known from Morrowind: Mod Loading Order (later: MLO) This would ensure that if your mod loading order is correct, first one will be picked up. I suggest having new key, like "OV%_" to make prioritised and overwriting different behaviours, as it may be useful. If "OV%_" is used, listed backgrounds/sounds will also check which mod loaded first and limit the list only to those entries. Additional work you will need to do here is to make `Mods Screen` to be able to change MLO in-game, so they can be listed in correct order. Additional optional `info.json` value will not hurt either, and can be good way of pre-determining MLO status for your mod, without manually setting it by player. ``` { "loading_number": -50 } ``` Example above would be run **before** anyone with number greater than -50.
1.0
Better "prioritised" system for panoramas and menu sounds & Mod Loading Order - Having possibility to prioritise panoramas and sounds is cool, as they will overwrite vanilla, but... They will not overwrite other mods. To explain it a bit in detail: if more than one mod use "PR%_" system, they will both shuffle through their data, so more than one mod can be prioritised. It's okay if you just want to overwrite vanilla, but not if you want to rule over all possible mods. Solution for this can be kinda simple, kinda not, mechanic known from Morrowind: Mod Loading Order (later: MLO) This would ensure that if your mod loading order is correct, first one will be picked up. I suggest having new key, like "OV%_" to make prioritised and overwriting different behaviours, as it may be useful. If "OV%_" is used, listed backgrounds/sounds will also check which mod loaded first and limit the list only to those entries. Additional work you will need to do here is to make `Mods Screen` to be able to change MLO in-game, so they can be listed in correct order. Additional optional `info.json` value will not hurt either, and can be good way of pre-determining MLO status for your mod, without manually setting it by player. ``` { "loading_number": -50 } ``` Example above would be run **before** anyone with number greater than -50.
non_defect
better prioritised system for panoramas and menu sounds mod loading order having possibility to prioritise panoramas and sounds is cool as they will overwrite vanilla but they will not overwrite other mods to explain it a bit in detail if more than one mod use pr system they will both shuffle through their data so more than one mod can be prioritised it s okay if you just want to overwrite vanilla but not if you want to rule over all possible mods solution for this can be kinda simple kinda not mechanic known from morrowind mod loading order later mlo this would ensure that if your mod loading order is correct first one will be picked up i suggest having new key like ov to make prioritised and overwriting different behaviours as it may be useful if ov is used listed backgrounds sounds will also check which mod loaded first and limit the list only to those entries additional work you will need to do here is to make mods screen to be able to change mlo in game so they can be listed in correct order additional optional info json value will not hurt either and can be good way of pre determining mlo status for your mod without manually setting it by player loading number example above would be run before anyone with number greater than
0
310,042
9,484,953,653
IssuesEvent
2019-04-22 08:36:59
facebook/create-react-app
https://api.github.com/repos/facebook/create-react-app
closed
Absolute URLS
priority: low (ignored issue template) stale
I would like to be able to use an absolute url in my index.html file but everytime i build it nothing changes no matter what i try in the package json even changing things in the webpack.config.prod.js E.G. not /static/js/somejsfile.js but https://hosturl.com/static/js/somejsfile.js
1.0
Absolute URLS - I would like to be able to use an absolute url in my index.html file but everytime i build it nothing changes no matter what i try in the package json even changing things in the webpack.config.prod.js E.G. not /static/js/somejsfile.js but https://hosturl.com/static/js/somejsfile.js
non_defect
absolute urls i would like to be able to use an absolute url in my index html file but everytime i build it nothing changes no matter what i try in the package json even changing things in the webpack config prod js e g not static js somejsfile js but
0
28,173
11,597,636,675
IssuesEvent
2020-02-24 21:18:04
gate5/struts-2.3.20
https://api.github.com/repos/gate5/struts-2.3.20
opened
WS-2019-0367 (Medium) detected in angular-1.2.23.js, angular-1.2.23.min.js
security vulnerability
## WS-2019-0367 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.23.js</b>, <b>angular-1.2.23.min.js</b></p></summary> <p> <details><summary><b>angular-1.2.23.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.js</a></p> <p>Path to vulnerable library: /struts-2.3.20/archetypes/struts2-archetype-angularjs/src/main/resources/archetype-resources/src/main/webapp/js/lib/angular/angular.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.23.js** (Vulnerable Library) </details> <details><summary><b>angular-1.2.23.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.min.js</a></p> <p>Path to vulnerable library: /struts-2.3.20/archetypes/struts2-archetype-angularjs/src/main/resources/archetype-resources/src/main/webapp/js/lib/angular/angular.min.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.23.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gate5/struts-2.3.20/commit/5e7aecd53f7232d1807fe90a7aed2d4ed0bd95b5">5e7aecd53f7232d1807fe90a7aed2d4ed0bd95b5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability found in Angular before 1.7.9. <p>Publish Date: 2020-01-08 <p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0367</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19">https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19</a></p> <p>Release Date: 2020-01-08</p> <p>Fix Resolution: angular - 1.7.9</p> </p> </details> <p></p>
True
WS-2019-0367 (Medium) detected in angular-1.2.23.js, angular-1.2.23.min.js - ## WS-2019-0367 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.23.js</b>, <b>angular-1.2.23.min.js</b></p></summary> <p> <details><summary><b>angular-1.2.23.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.js</a></p> <p>Path to vulnerable library: /struts-2.3.20/archetypes/struts2-archetype-angularjs/src/main/resources/archetype-resources/src/main/webapp/js/lib/angular/angular.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.23.js** (Vulnerable Library) </details> <details><summary><b>angular-1.2.23.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.23/angular.min.js</a></p> <p>Path to vulnerable library: /struts-2.3.20/archetypes/struts2-archetype-angularjs/src/main/resources/archetype-resources/src/main/webapp/js/lib/angular/angular.min.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.23.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gate5/struts-2.3.20/commit/5e7aecd53f7232d1807fe90a7aed2d4ed0bd95b5">5e7aecd53f7232d1807fe90a7aed2d4ed0bd95b5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability found in Angular before 1.7.9. <p>Publish Date: 2020-01-08 <p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0367</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19">https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19</a></p> <p>Release Date: 2020-01-08</p> <p>Fix Resolution: angular - 1.7.9</p> </p> </details> <p></p>
non_defect
ws medium detected in angular js angular min js ws medium severity vulnerability vulnerable libraries angular js angular min js angular js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library struts archetypes archetype angularjs src main resources archetype resources src main webapp js lib angular angular js dependency hierarchy x angular js vulnerable library angular min js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library struts archetypes archetype angularjs src main resources archetype resources src main webapp js lib angular angular min js dependency hierarchy x angular min js vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability found in angular before publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution angular
0
59,150
17,016,088,079
IssuesEvent
2021-07-02 12:15:53
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
DatabaseWarning too OSM-specific, impossible to disable ("changes are saved to the main OpenStreetMap database")
Component: potlatch2 Priority: minor Type: defect
**[Submitted to the original trac issue database at 9.59pm, Thursday, 13th September 2012]** The save dialog (```SaveDialog.mxml```) contains a hardcoded warning (databaseWarning): ``` Your changes are saved to the main OpenStreetMap database for everyone to see. This is not your own private map. Please don&apos;t save test changes or edits for personal use only. ``` I couldn't find any way to disable or customize this. When using Potlatch as an editor for custom instances, this warning is extremely confusing. Additionally, it seems that this warning is not localized. Please allow some easy way (e.g. a flash var) to customize/disable this warning. As the coding guidelines say "Both Potlatch 2 and Halcyon can be deployed locally (i.e. not on openstreetmap.org), and indeed can talk to databases other than the main OSM instance. So don't be too OSM-specific", I am labeling this as defect.
1.0
DatabaseWarning too OSM-specific, impossible to disable ("changes are saved to the main OpenStreetMap database") - **[Submitted to the original trac issue database at 9.59pm, Thursday, 13th September 2012]** The save dialog (```SaveDialog.mxml```) contains a hardcoded warning (databaseWarning): ``` Your changes are saved to the main OpenStreetMap database for everyone to see. This is not your own private map. Please don&apos;t save test changes or edits for personal use only. ``` I couldn't find any way to disable or customize this. When using Potlatch as an editor for custom instances, this warning is extremely confusing. Additionally, it seems that this warning is not localized. Please allow some easy way (e.g. a flash var) to customize/disable this warning. As the coding guidelines say "Both Potlatch 2 and Halcyon can be deployed locally (i.e. not on openstreetmap.org), and indeed can talk to databases other than the main OSM instance. So don't be too OSM-specific", I am labeling this as defect.
defect
databasewarning too osm specific impossible to disable changes are saved to the main openstreetmap database the save dialog savedialog mxml contains a hardcoded warning databasewarning your changes are saved to the main openstreetmap database for everyone to see this is not your own private map please don apos t save test changes or edits for personal use only i couldn t find any way to disable or customize this when using potlatch as an editor for custom instances this warning is extremely confusing additionally it seems that this warning is not localized please allow some easy way e g a flash var to customize disable this warning as the coding guidelines say both potlatch and halcyon can be deployed locally i e not on openstreetmap org and indeed can talk to databases other than the main osm instance so don t be too osm specific i am labeling this as defect
1
33,880
2,773,273,743
IssuesEvent
2015-05-03 13:54:42
web2py/web2py
https://api.github.com/repos/web2py/web2py
closed
Import/Export CSV Broken On Complex Queries (Admin - Database Admin - Table - Query)
bug imported Priority-Medium
_From [grant.tr..._at_gmail.com](https://code.google.com/u/112191023930957557807/) on March 22, 2014 23:40:41_ 1. Login to admin interface 2. Select "Edit Application" 3. Select "Database Administration" 4. Select a table to inspect 5. Default query has an "Import/Export" section for importing / exporting CSV - GOOD 6. The existing query looks like 'db.auth_user.id>0' - OK 7. Now put parenthesis around the query '(db.auth_user.id>0)' - BAD 8. "export as csv" no longer works - BAD 9. Import is missing entirely - BAD This also happens for more complex queries. Version information reported by web2py: 2.9.5-stable+timestamp.2014.03.16.02.35.39 (Running on Apache/2.2.22 (Ubuntu), Python 2.7.3) _Original issue: http://code.google.com/p/web2py/issues/detail?id=1907_
1.0
Import/Export CSV Broken On Complex Queries (Admin - Database Admin - Table - Query) - _From [grant.tr..._at_gmail.com](https://code.google.com/u/112191023930957557807/) on March 22, 2014 23:40:41_ 1. Login to admin interface 2. Select "Edit Application" 3. Select "Database Administration" 4. Select a table to inspect 5. Default query has an "Import/Export" section for importing / exporting CSV - GOOD 6. The existing query looks like 'db.auth_user.id>0' - OK 7. Now put parenthesis around the query '(db.auth_user.id>0)' - BAD 8. "export as csv" no longer works - BAD 9. Import is missing entirely - BAD This also happens for more complex queries. Version information reported by web2py: 2.9.5-stable+timestamp.2014.03.16.02.35.39 (Running on Apache/2.2.22 (Ubuntu), Python 2.7.3) _Original issue: http://code.google.com/p/web2py/issues/detail?id=1907_
non_defect
import export csv broken on complex queries admin database admin table query from on march login to admin interface select edit application select database administration select a table to inspect default query has an import export section for importing exporting csv good the existing query looks like db auth user id ok now put parenthesis around the query db auth user id bad export as csv no longer works bad import is missing entirely bad this also happens for more complex queries version information reported by stable timestamp running on apache ubuntu python original issue
0
33,438
7,123,672,949
IssuesEvent
2018-01-19 16:08:18
contao/core-bundle
https://api.github.com/repos/contao/core-bundle
opened
Update to Symfony 3.4
defect
Now that the Symfony LTS version has been released, we should allow it in the `composer.json` file of our bundles. The version constraint should be one of 1. `^3.4` 2. `^3.3|^3.4` I think I would prefer no. 1, because Symfony 3.3 is no longer maintained. On the other hand, there have been some changes in Symfony 3.4 which might not be 100% backwards compatible. @contao/developers Which version constraint would you prefer?
1.0
Update to Symfony 3.4 - Now that the Symfony LTS version has been released, we should allow it in the `composer.json` file of our bundles. The version constraint should be one of 1. `^3.4` 2. `^3.3|^3.4` I think I would prefer no. 1, because Symfony 3.3 is no longer maintained. On the other hand, there have been some changes in Symfony 3.4 which might not be 100% backwards compatible. @contao/developers Which version constraint would you prefer?
defect
update to symfony now that the symfony lts version has been released we should allow it in the composer json file of our bundles the version constraint should be one of i think i would prefer no because symfony is no longer maintained on the other hand there have been some changes in symfony which might not be backwards compatible contao developers which version constraint would you prefer
1
54,673
13,877,055,240
IssuesEvent
2020-10-17 02:02:29
AeroScripts/QuestieDev
https://api.github.com/repos/AeroScripts/QuestieDev
closed
Waypoints for objectives are missing
Questie - Map / Minimap Type - Defect
## Bug description Waypoints are shown for quest givers and finishers, but not for objective NPCs (e.g. Defias Messenger in Westfall). ## Questie version All versions since introduction of the waypoint feature are affected.
1.0
Waypoints for objectives are missing - ## Bug description Waypoints are shown for quest givers and finishers, but not for objective NPCs (e.g. Defias Messenger in Westfall). ## Questie version All versions since introduction of the waypoint feature are affected.
defect
waypoints for objectives are missing bug description waypoints are shown for quest givers and finishers but not for objective npcs e g defias messenger in westfall questie version all versions since introduction of the waypoint feature are affected
1
7,703
2,610,433,837
IssuesEvent
2015-02-26 20:22:08
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
FF 29.0.1, scribefire add-on, and zemanta
auto-migrated Priority-Medium Type-Defect
``` What's the problem? I have FF 29.0.1, and a Blogger blog at http://drwebworld.blogspot.com/ and add-ons for ScribeFire and Zemanta. I use BlogThis! I make a temp entry, using BlogThis! Save as draft. Then, I try and open that draft post in ScribeFire. It opens when I pick it from my list of articles. BUT, the problem occurs now. There is no linkage, or buttons, or intergration of Zemanta. It's just not there. Screenshot attached of what I see. What browser are you using? FF 29.0.1 What Operating system are you using Windows 7 Enterprise What version of ScribeFire are you running? latest...4.3.1 What Blog Type are you having this problem with? Please include version # if known or applicable Blogger ``` ----- Original issue reported on code.google.com by `drw...@gmail.com` on 2 Jun 2014 at 10:24 Attachments: * [ScribeFire.jpg](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-839/comment-0/ScribeFire.jpg)
1.0
FF 29.0.1, scribefire add-on, and zemanta - ``` What's the problem? I have FF 29.0.1, and a Blogger blog at http://drwebworld.blogspot.com/ and add-ons for ScribeFire and Zemanta. I use BlogThis! I make a temp entry, using BlogThis! Save as draft. Then, I try and open that draft post in ScribeFire. It opens when I pick it from my list of articles. BUT, the problem occurs now. There is no linkage, or buttons, or intergration of Zemanta. It's just not there. Screenshot attached of what I see. What browser are you using? FF 29.0.1 What Operating system are you using Windows 7 Enterprise What version of ScribeFire are you running? latest...4.3.1 What Blog Type are you having this problem with? Please include version # if known or applicable Blogger ``` ----- Original issue reported on code.google.com by `drw...@gmail.com` on 2 Jun 2014 at 10:24 Attachments: * [ScribeFire.jpg](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-839/comment-0/ScribeFire.jpg)
defect
ff scribefire add on and zemanta what s the problem i have ff and a blogger blog at and add ons for scribefire and zemanta i use blogthis i make a temp entry using blogthis save as draft then i try and open that draft post in scribefire it opens when i pick it from my list of articles but the problem occurs now there is no linkage or buttons or intergration of zemanta it s just not there screenshot attached of what i see what browser are you using ff what operating system are you using windows enterprise what version of scribefire are you running latest what blog type are you having this problem with please include version if known or applicable blogger original issue reported on code google com by drw gmail com on jun at attachments
1
146,667
5,625,815,629
IssuesEvent
2017-04-04 20:19:41
SCIInstitute/ALMA-TDA
https://api.github.com/repos/SCIInstitute/ALMA-TDA
closed
command line interface
enhancement high priority
need to be able to process cubes from the command line to fit into existing workflows.
1.0
command line interface - need to be able to process cubes from the command line to fit into existing workflows.
non_defect
command line interface need to be able to process cubes from the command line to fit into existing workflows
0
252,491
8,036,981,520
IssuesEvent
2018-07-30 10:59:20
reconhub/epiflows
https://api.github.com/repos/reconhub/epiflows
closed
Simplify arguments of estimate_risk_spread
discussion enhancement hackathon priority
The function currently has many arguments: https://github.com/reconhub/epiflows/blob/4ccd83e9f07aa5228c47abf9858c3656aea13131/R/estimate_risk_spread.R#L42-L55 If possible, simplify the arguments, or put everything in a list which has defaults and can be checked, a la `outbreaker2`: https://github.com/reconhub/outbreaker2/blob/master/R/create_config.R
1.0
Simplify arguments of estimate_risk_spread - The function currently has many arguments: https://github.com/reconhub/epiflows/blob/4ccd83e9f07aa5228c47abf9858c3656aea13131/R/estimate_risk_spread.R#L42-L55 If possible, simplify the arguments, or put everything in a list which has defaults and can be checked, a la `outbreaker2`: https://github.com/reconhub/outbreaker2/blob/master/R/create_config.R
non_defect
simplify arguments of estimate risk spread the function currently has many arguments if possible simplify the arguments or put everything in a list which has defaults and can be checked a la
0
227,129
17,374,957,141
IssuesEvent
2021-07-30 19:27:56
open-contracting/standard
https://api.github.com/repos/open-contracting/standard
closed
Worked example: Move discussion of contracting process out of unsuccessful tenders
Focus - Documentation
The worked example was merged in #1007 based on [this outline](https://docs.google.com/document/d/1bIOMCC1sQjrz2UO5bkc2PcHtTaDA2C1W7jeXRRqGjfE/edit). However, it spends the first half explaining what a contracting process is. This content should be moved/removed, since #866 will define a contracting process.
1.0
Worked example: Move discussion of contracting process out of unsuccessful tenders - The worked example was merged in #1007 based on [this outline](https://docs.google.com/document/d/1bIOMCC1sQjrz2UO5bkc2PcHtTaDA2C1W7jeXRRqGjfE/edit). However, it spends the first half explaining what a contracting process is. This content should be moved/removed, since #866 will define a contracting process.
non_defect
worked example move discussion of contracting process out of unsuccessful tenders the worked example was merged in based on however it spends the first half explaining what a contracting process is this content should be moved removed since will define a contracting process
0
6,790
2,860,130,830
IssuesEvent
2015-06-03 14:31:33
AppStateESS/homestead
https://api.github.com/repos/AppStateESS/homestead
closed
Add smoking preference to application
Applications enhancement NeedsTesting PendingRelease
Add a smoking preference to the freshmen applications. Use it for matching in the auto assigner.
1.0
Add smoking preference to application - Add a smoking preference to the freshmen applications. Use it for matching in the auto assigner.
non_defect
add smoking preference to application add a smoking preference to the freshmen applications use it for matching in the auto assigner
0
9,682
2,615,165,787
IssuesEvent
2015-03-01 06:46:07
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Waiting for beacon from, Receive timeout occurred, WPS transaction failed
auto-migrated Priority-Triage Type-Defect
``` hey, i have my drivers ath9k, but when i use reaver just stuck on [+] Waiting for beacon from , and do nothing. When i use reaver -i mon0 -b ********** -vv -S -N -L its say: [+] Waiting for beacon from ************* [+] Switching mon0 to channel 11 [+] Associated with ************** (ESSID: CablemasF949) [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [!] WARNING: Receive timeout occurred [+] Sending WSC NACK [!] WPS transaction failed (code: 0x02), re-trying last pin And the same..... Look, i am noob for all this and i dont knonw what i do wrong, so please explain me how to make reaver work? __________________________________________________________________________ 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) REAVER 1.4 1. What operating system are you using (Linux is the only supported OS)? UBUNTU 12.10 2. Is your wireless card in monitor mode (yes/no)? YES 3. What is the signal strength of the Access Point you are trying to crack? -73 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b ****** 6. Please describe what you think the issue is. DONT KNOW 7. Paste the output from Reaver below. [+] Waiting for beacon from ***** [+] Associated with ******* ``` Original issue reported on code.google.com by `solloman...@gmail.com` on 2 Apr 2013 at 4:46
1.0
Waiting for beacon from, Receive timeout occurred, WPS transaction failed - ``` hey, i have my drivers ath9k, but when i use reaver just stuck on [+] Waiting for beacon from , and do nothing. When i use reaver -i mon0 -b ********** -vv -S -N -L its say: [+] Waiting for beacon from ************* [+] Switching mon0 to channel 11 [+] Associated with ************** (ESSID: CablemasF949) [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [!] WARNING: Receive timeout occurred [+] Sending WSC NACK [!] WPS transaction failed (code: 0x02), re-trying last pin And the same..... Look, i am noob for all this and i dont knonw what i do wrong, so please explain me how to make reaver work? __________________________________________________________________________ 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) REAVER 1.4 1. What operating system are you using (Linux is the only supported OS)? UBUNTU 12.10 2. Is your wireless card in monitor mode (yes/no)? YES 3. What is the signal strength of the Access Point you are trying to crack? -73 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b ****** 6. Please describe what you think the issue is. DONT KNOW 7. Paste the output from Reaver below. [+] Waiting for beacon from ***** [+] Associated with ******* ``` Original issue reported on code.google.com by `solloman...@gmail.com` on 2 Apr 2013 at 4:46
defect
waiting for beacon from receive timeout occurred wps transaction failed hey i have my drivers but when i use reaver just stuck on waiting for beacon from and do nothing when i use reaver i b vv s n l its say waiting for beacon from switching to channel associated with essid trying pin sending eapol start request received identity request sending identity response warning receive timeout occurred sending wsc nack wps transaction failed code re trying last pin and the same look i am noob for all this and i dont knonw what i do wrong so please explain me how to make reaver work what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os ubuntu is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the entire command line string you are supplying to reaver reaver i b please describe what you think the issue is dont know paste the output from reaver below waiting for beacon from associated with original issue reported on code google com by solloman gmail com on apr at
1
2,978
2,607,968,279
IssuesEvent
2015-02-26 00:43:22
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
opened
[Feature Required] Compaction read/write speed limit
auto-migrated Priority-Medium Type-Defect
``` We are running SSDB(a leveldb server, https://github.com/ideawu/ssdb), with 28G data total, 400k updates per day. LevelDB compaction happends 6-10 times per day. While compaction is in progress, all reads and writes made to leveldb is significant slow. The disk IO is almost full, this is the reason(I think) why leveldb slows down. What is the expected output? What do you see instead? Compaction won't slow down normal operations, leveldb's compaction thread should limit the disk read/write speed. What version of the product are you using? On what operating system? 1.9.0 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `wuzuy...@gmail.com` on 14 May 2013 at 3:52
1.0
[Feature Required] Compaction read/write speed limit - ``` We are running SSDB(a leveldb server, https://github.com/ideawu/ssdb), with 28G data total, 400k updates per day. LevelDB compaction happends 6-10 times per day. While compaction is in progress, all reads and writes made to leveldb is significant slow. The disk IO is almost full, this is the reason(I think) why leveldb slows down. What is the expected output? What do you see instead? Compaction won't slow down normal operations, leveldb's compaction thread should limit the disk read/write speed. What version of the product are you using? On what operating system? 1.9.0 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `wuzuy...@gmail.com` on 14 May 2013 at 3:52
defect
compaction read write speed limit we are running ssdb a leveldb server with data total updates per day leveldb compaction happends times per day while compaction is in progress all reads and writes made to leveldb is significant slow the disk io is almost full this is the reason i think why leveldb slows down what is the expected output what do you see instead compaction won t slow down normal operations leveldb s compaction thread should limit the disk read write speed what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by wuzuy gmail com on may at
1
203,856
15,392,427,810
IssuesEvent
2021-03-03 15:37:12
firebase/firebase-ios-sdk
https://api.github.com/repos/firebase/firebase-ios-sdk
closed
Nightly Testing Report For Presubmit Testing
nightly-testing-report-generation-test
# Nightly Testing Report For Presubmit Testing This issue([log](https://github.com/firebase/firebase-ios-sdk/actions/runs/616153570)) is generated at 03/02/2021 18:31 PM, fetching workflow runs triggered in the last 24 hrs secs. | Workflow |03/01| | -------- | -------- | | [firestore](https://github.com/firebase/firebase-ios-sdk/blob/master/.github/workflows/firestore.yml) | [failure](https://github.com/firebase/firebase-ios-sdk/actions/runs/612949071) | | [mlmodeldownloader](https://github.com/firebase/firebase-ios-sdk/blob/master/.github/workflows/mlmodeldownloader.yml) | [failure](https://github.com/firebase/firebase-ios-sdk/actions/runs/613185286) |
2.0
Nightly Testing Report For Presubmit Testing - # Nightly Testing Report For Presubmit Testing This issue([log](https://github.com/firebase/firebase-ios-sdk/actions/runs/616153570)) is generated at 03/02/2021 18:31 PM, fetching workflow runs triggered in the last 24 hrs secs. | Workflow |03/01| | -------- | -------- | | [firestore](https://github.com/firebase/firebase-ios-sdk/blob/master/.github/workflows/firestore.yml) | [failure](https://github.com/firebase/firebase-ios-sdk/actions/runs/612949071) | | [mlmodeldownloader](https://github.com/firebase/firebase-ios-sdk/blob/master/.github/workflows/mlmodeldownloader.yml) | [failure](https://github.com/firebase/firebase-ios-sdk/actions/runs/613185286) |
non_defect
nightly testing report for presubmit testing nightly testing report for presubmit testing this issue is generated at pm fetching workflow runs triggered in the last hrs secs workflow
0
198,177
22,617,949,315
IssuesEvent
2022-06-30 01:26:04
YauheniPo/mix_api_-_selenium_test_framework
https://api.github.com/repos/YauheniPo/mix_api_-_selenium_test_framework
opened
CVE-2022-33879 (Medium) detected in tika-core-1.20.jar
security vulnerability
## CVE-2022-33879 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.20.jar</b></p></summary> <p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also includes the core facades for the Tika API.</p> <p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tika/tika-core/1.20/tika-core-1.20.jar</p> <p> Dependency Hierarchy: - allure-testng-2.13.1.jar (Root Library) - allure-java-commons-2.13.1.jar - :x: **tika-core-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YauheniPo/mix_api_-_selenium_test_framework/commit/071d424c639b56ed58d78122a50c578a7055ec90">071d424c639b56ed58d78122a50c578a7055ec90</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The initial fixes in CVE-2022-30126 and CVE-2022-30973 for regexes in the StandardsExtractingContentHandler were insufficient, and we found a separate, new regex DoS in a different regex in the StandardsExtractingContentHandler. These are now fixed in 1.28.4 and 2.4.1. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33879>CVE-2022-33879</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33879</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution (org.apache.tika:tika-core): 2.0.0-ALPHA</p> <p>Direct dependency fix Resolution (io.qameta.allure:allure-testng): 2.15.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-33879 (Medium) detected in tika-core-1.20.jar - ## CVE-2022-33879 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.20.jar</b></p></summary> <p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also includes the core facades for the Tika API.</p> <p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tika/tika-core/1.20/tika-core-1.20.jar</p> <p> Dependency Hierarchy: - allure-testng-2.13.1.jar (Root Library) - allure-java-commons-2.13.1.jar - :x: **tika-core-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YauheniPo/mix_api_-_selenium_test_framework/commit/071d424c639b56ed58d78122a50c578a7055ec90">071d424c639b56ed58d78122a50c578a7055ec90</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The initial fixes in CVE-2022-30126 and CVE-2022-30973 for regexes in the StandardsExtractingContentHandler were insufficient, and we found a separate, new regex DoS in a different regex in the StandardsExtractingContentHandler. These are now fixed in 1.28.4 and 2.4.1. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33879>CVE-2022-33879</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33879</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution (org.apache.tika:tika-core): 2.0.0-ALPHA</p> <p>Direct dependency fix Resolution (io.qameta.allure:allure-testng): 2.15.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in tika core jar cve medium severity vulnerability vulnerable library tika core jar this is the core apache tika™ toolkit library from which all other modules inherit functionality it also includes the core facades for the tika api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org apache tika tika core tika core jar dependency hierarchy allure testng jar root library allure java commons jar x tika core jar vulnerable library found in head commit a href found in base branch master vulnerability details the initial fixes in cve and cve for regexes in the standardsextractingcontenthandler were insufficient and we found a separate new regex dos in a different regex in the standardsextractingcontenthandler these are now fixed in and publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tika tika core alpha direct dependency fix resolution io qameta allure allure testng step up your open source security game with mend
0
70,000
22,781,985,444
IssuesEvent
2022-07-08 21:01:58
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
upgrading to v1.11.0 invalidates session
T-Defect
### Steps to reproduce 1. Have existing installation and signed in profile of electron-desktop v1.10.15. 2. Install v1.11.0 3. Start element-desktop ### Outcome #### What did you expect? element-desktop starts and shows a signed-in view with rooms just as on v1.10.15. #### What happened instead? The following modal is displayed with a "Sign out" button being the only actionable: ``` Missing session data Some session data, including encrypted message keys, is missing. Sign out and sign in to fix this, restoring keys from backup. Your browser likely removed this data when running low on disk space. To help us prevent this in future, please send us logs. ``` ![element-unexplained-sign-out](https://user-images.githubusercontent.com/74199244/177890678-2e832830-1a00-4309-ae55-8c6ff11b99e9.png) ### Operating system Linux (NixOS) ### Application version 1.11.0 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
upgrading to v1.11.0 invalidates session - ### Steps to reproduce 1. Have existing installation and signed in profile of electron-desktop v1.10.15. 2. Install v1.11.0 3. Start element-desktop ### Outcome #### What did you expect? element-desktop starts and shows a signed-in view with rooms just as on v1.10.15. #### What happened instead? The following modal is displayed with a "Sign out" button being the only actionable: ``` Missing session data Some session data, including encrypted message keys, is missing. Sign out and sign in to fix this, restoring keys from backup. Your browser likely removed this data when running low on disk space. To help us prevent this in future, please send us logs. ``` ![element-unexplained-sign-out](https://user-images.githubusercontent.com/74199244/177890678-2e832830-1a00-4309-ae55-8c6ff11b99e9.png) ### Operating system Linux (NixOS) ### Application version 1.11.0 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
upgrading to invalidates session steps to reproduce have existing installation and signed in profile of electron desktop install start element desktop outcome what did you expect element desktop starts and shows a signed in view with rooms just as on what happened instead the following modal is displayed with a sign out button being the only actionable missing session data some session data including encrypted message keys is missing sign out and sign in to fix this restoring keys from backup your browser likely removed this data when running low on disk space to help us prevent this in future please send us logs operating system linux nixos application version how did you install the app no response homeserver no response will you send logs no
1
10,824
3,436,612,409
IssuesEvent
2015-12-12 14:55:36
lintool/warcbase
https://api.github.com/repos/lintool/warcbase
opened
Translate `DetectLanguage` pig script into Scala; Incorporate into RecordRDD?
documentation feature
Under pig, we had this script: ``` register 'target/warcbase-0.1.0-SNAPSHOT-fatjar.jar'; DEFINE ArcLoader org.warcbase.pig.ArcLoader(); DEFINE DetectLanguage org.warcbase.pig.piggybank.DetectLanguage(); DEFINE ExtractRawText org.warcbase.pig.piggybank.ExtractRawText(); DEFINE ExtractTopLevelDomain org.warcbase.pig.piggybank.ExtractTopLevelDomain(); raw = load '/shared/collections/CanadianPoliticalParties/arc/' using ArcLoader as (url: chararray, date: chararray, mime: chararray, content: bytearray); a = filter raw by mime == 'text/html' and date is not null; b = foreach a generate SUBSTRING(date, 0, 6) as date, REPLACE(ExtractTopLevelDomain(url), '^\\s*www\\.', '') as url, content; c = filter b by url == 'greenparty.ca'; d = foreach c generate date, url, ExtractRawText((chararray) content) as text; e = foreach d generate date, url, DetectLanguage(text) as lang, text; store e into 'cpp.text-greenparty'; ``` Could/should we bake our `DetectLanguage` function into `RecordRDD.scala` - i.e. alongside `keepMimeTypes` or `keepDomains`? That would fit best with our plain text scripts, so for example this could work (added a fictional line after `keepDomains` and before the `map` function. ``` import org.warcbase.spark.matchbox.{RemoveHTML, RecordLoader} import org.warcbase.spark.rdd.RecordRDD._ RecordLoader.loadArc("src/test/resources/arc/example.arc.gz", sc) .keepValidPages() .keepDomains(Set("greenparty.ca")) .keepLanguages(Set("en")) .map(r => (r.getCrawldate, r.getDomain, r.getUrl, RemoveHTML(r.getContentString))) .saveAsTextFile("out/") ``` I created a new branch, keep-languages, but realize I should check if this is feasible first (since we have largely used `Record.RDD` for URL and mime-type filtering. If anybody has time or ability to tackle this, please use that branch.
1.0
Translate `DetectLanguage` pig script into Scala; Incorporate into RecordRDD? - Under pig, we had this script: ``` register 'target/warcbase-0.1.0-SNAPSHOT-fatjar.jar'; DEFINE ArcLoader org.warcbase.pig.ArcLoader(); DEFINE DetectLanguage org.warcbase.pig.piggybank.DetectLanguage(); DEFINE ExtractRawText org.warcbase.pig.piggybank.ExtractRawText(); DEFINE ExtractTopLevelDomain org.warcbase.pig.piggybank.ExtractTopLevelDomain(); raw = load '/shared/collections/CanadianPoliticalParties/arc/' using ArcLoader as (url: chararray, date: chararray, mime: chararray, content: bytearray); a = filter raw by mime == 'text/html' and date is not null; b = foreach a generate SUBSTRING(date, 0, 6) as date, REPLACE(ExtractTopLevelDomain(url), '^\\s*www\\.', '') as url, content; c = filter b by url == 'greenparty.ca'; d = foreach c generate date, url, ExtractRawText((chararray) content) as text; e = foreach d generate date, url, DetectLanguage(text) as lang, text; store e into 'cpp.text-greenparty'; ``` Could/should we bake our `DetectLanguage` function into `RecordRDD.scala` - i.e. alongside `keepMimeTypes` or `keepDomains`? That would fit best with our plain text scripts, so for example this could work (added a fictional line after `keepDomains` and before the `map` function. ``` import org.warcbase.spark.matchbox.{RemoveHTML, RecordLoader} import org.warcbase.spark.rdd.RecordRDD._ RecordLoader.loadArc("src/test/resources/arc/example.arc.gz", sc) .keepValidPages() .keepDomains(Set("greenparty.ca")) .keepLanguages(Set("en")) .map(r => (r.getCrawldate, r.getDomain, r.getUrl, RemoveHTML(r.getContentString))) .saveAsTextFile("out/") ``` I created a new branch, keep-languages, but realize I should check if this is feasible first (since we have largely used `Record.RDD` for URL and mime-type filtering. If anybody has time or ability to tackle this, please use that branch.
non_defect
translate detectlanguage pig script into scala incorporate into recordrdd under pig we had this script register target warcbase snapshot fatjar jar define arcloader org warcbase pig arcloader define detectlanguage org warcbase pig piggybank detectlanguage define extractrawtext org warcbase pig piggybank extractrawtext define extracttopleveldomain org warcbase pig piggybank extracttopleveldomain raw load shared collections canadianpoliticalparties arc using arcloader as url chararray date chararray mime chararray content bytearray a filter raw by mime text html and date is not null b foreach a generate substring date as date replace extracttopleveldomain url s www as url content c filter b by url greenparty ca d foreach c generate date url extractrawtext chararray content as text e foreach d generate date url detectlanguage text as lang text store e into cpp text greenparty could should we bake our detectlanguage function into recordrdd scala i e alongside keepmimetypes or keepdomains that would fit best with our plain text scripts so for example this could work added a fictional line after keepdomains and before the map function import org warcbase spark matchbox removehtml recordloader import org warcbase spark rdd recordrdd recordloader loadarc src test resources arc example arc gz sc keepvalidpages keepdomains set greenparty ca keeplanguages set en map r r getcrawldate r getdomain r geturl removehtml r getcontentstring saveastextfile out i created a new branch keep languages but realize i should check if this is feasible first since we have largely used record rdd for url and mime type filtering if anybody has time or ability to tackle this please use that branch
0
20,181
3,309,982,771
IssuesEvent
2015-11-05 05:33:32
navjagpal/browser-timetracker
https://api.github.com/repos/navjagpal/browser-timetracker
closed
Use the Idle API instead of mouse listeners and content scripts.
auto-migrated Priority-Medium Type-Defect
``` Use the Idle API http://code.google.com/chrome/extensions/idle.html Instead of placing mouse listeners all over the DOM, it is more efficient to do that using the built in API. ``` Original issue reported on code.google.com by `m0.inter...@gmail.com` on 9 Jul 2011 at 2:24
1.0
Use the Idle API instead of mouse listeners and content scripts. - ``` Use the Idle API http://code.google.com/chrome/extensions/idle.html Instead of placing mouse listeners all over the DOM, it is more efficient to do that using the built in API. ``` Original issue reported on code.google.com by `m0.inter...@gmail.com` on 9 Jul 2011 at 2:24
defect
use the idle api instead of mouse listeners and content scripts use the idle api instead of placing mouse listeners all over the dom it is more efficient to do that using the built in api original issue reported on code google com by inter gmail com on jul at
1
16,249
10,695,102,272
IssuesEvent
2019-10-23 12:21:29
virtualsatellite/VirtualSatellite4-Core
https://api.github.com/repos/virtualsatellite/VirtualSatellite4-Core
closed
Add section for CATIA import / export in User Guide
comfort/usability quality
The user guide should explain how to use the CATIA import / export functionality. It should also explain how to setup the Visual Basic scripts in CATIA.
True
Add section for CATIA import / export in User Guide - The user guide should explain how to use the CATIA import / export functionality. It should also explain how to setup the Visual Basic scripts in CATIA.
non_defect
add section for catia import export in user guide the user guide should explain how to use the catia import export functionality it should also explain how to setup the visual basic scripts in catia
0
42,620
12,918,065,565
IssuesEvent
2020-07-16 01:41:27
rvvergara/bookstore-react-redux-app
https://api.github.com/repos/rvvergara/bookstore-react-redux-app
opened
CVE-2018-19797 (Medium) detected in node-sass-4.14.1.tgz, node-sass-v4.13.1
security vulnerability
## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/bookstore-react-redux-app/package.json</p> <p>Path to vulnerable library: /bookstore-react-redux-app/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-react-redux-app/commit/164012cab19f3382d89ee6a3e501e0356df2e377">164012cab19f3382d89ee6a3e501e0356df2e377</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass-3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19797 (Medium) detected in node-sass-4.14.1.tgz, node-sass-v4.13.1 - ## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/bookstore-react-redux-app/package.json</p> <p>Path to vulnerable library: /bookstore-react-redux-app/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-react-redux-app/commit/164012cab19f3382d89ee6a3e501e0356df2e377">164012cab19f3382d89ee6a3e501e0356df2e377</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass-3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in node sass tgz node sass cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm bookstore react redux app package json path to vulnerable library bookstore react redux app node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a null pointer dereference in the function sass selector list populate extends in sharedptr hpp used by ast cpp and ast selectors cpp may cause a denial of service application crash via a crafted sass input file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
57,615
15,882,891,658
IssuesEvent
2021-04-09 16:38:06
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Changing zvol_volmode to/from 3 doesn't hide/expose devices
Status: Triage Needed Type: Defect
### System information Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Buster Linux Kernel | Linux tarta 4.19.0-16-amd64 #1 SMP Debian 4.19.181-1 (2021-03-19) x86_64 GNU/Linux Architecture | amd64 ZFS Version | 2.0.3-1~bpo10+1 ### Describe the problem you're observing ### Describe how to reproduce the problem ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
Changing zvol_volmode to/from 3 doesn't hide/expose devices - ### System information Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Buster Linux Kernel | Linux tarta 4.19.0-16-amd64 #1 SMP Debian 4.19.181-1 (2021-03-19) x86_64 GNU/Linux Architecture | amd64 ZFS Version | 2.0.3-1~bpo10+1 ### Describe the problem you're observing ### Describe how to reproduce the problem ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
changing zvol volmode to from doesn t hide expose devices system information type version name distribution name debian distribution version buster linux kernel linux tarta smp debian gnu linux architecture zfs version describe the problem you re observing describe how to reproduce the problem include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
323,542
23,953,381,909
IssuesEvent
2022-09-12 13:18:41
rstudio/shinytest2
https://api.github.com/repos/rstudio/shinytest2
closed
Documentation: screenshot delay is seconds, not milliseconds
bug documentation
The documentation for `AppDriver$expect_screenshot()` and `AppDriver$get_screenshot()` says the `delay` argument is the delay in milliseconds, but the documentation for `chromote::ChromoteSession` says that the `delay` argument is in seconds, so the documentation for `shinytest2` should be corrected. I've checked and setting `delay = 200` delays for 200 seconds, not 200 milliseconds.
1.0
Documentation: screenshot delay is seconds, not milliseconds - The documentation for `AppDriver$expect_screenshot()` and `AppDriver$get_screenshot()` says the `delay` argument is the delay in milliseconds, but the documentation for `chromote::ChromoteSession` says that the `delay` argument is in seconds, so the documentation for `shinytest2` should be corrected. I've checked and setting `delay = 200` delays for 200 seconds, not 200 milliseconds.
non_defect
documentation screenshot delay is seconds not milliseconds the documentation for appdriver expect screenshot and appdriver get screenshot says the delay argument is the delay in milliseconds but the documentation for chromote chromotesession says that the delay argument is in seconds so the documentation for should be corrected i ve checked and setting delay delays for seconds not milliseconds
0
15,069
2,846,324,425
IssuesEvent
2015-05-29 10:41:08
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
checkbox label attribute glitch
Defect
What was done: ```php <?= $this->Form->input('active', ['type' => 'checkbox', 'label' => 'Activate']); ?> ``` Expected output: ```html <label for="active">Activate</label> <input type = 'checkbox' name="active" id="active" /> ``` Rendered output: ```html <label for='active'> <input type = 'checkbox' name="active" id="active" /> </label> ```
1.0
checkbox label attribute glitch - What was done: ```php <?= $this->Form->input('active', ['type' => 'checkbox', 'label' => 'Activate']); ?> ``` Expected output: ```html <label for="active">Activate</label> <input type = 'checkbox' name="active" id="active" /> ``` Rendered output: ```html <label for='active'> <input type = 'checkbox' name="active" id="active" /> </label> ```
defect
checkbox label attribute glitch what was done php form input active expected output html activate rendered output html
1
59,520
17,023,150,360
IssuesEvent
2021-07-03 00:36:04
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
API 0.3 traces display has slow query
Component: api Priority: major Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 6.42pm, Friday, 23rd March 2007]** found at source:/www.openstreetmap.org/ruby/api/osm/dao.rb@R1589#L431 seems hard to imagine that cross-reffing 20000 trace files and 5000 users could become a major drain on db, but that is apparently the situation. whatever the exact situation, the sql in here looks pretty grim. anyways, in the course of putting the traces stuff into 0.4 API i needed to check this code, so i thought i'd record results: without the test data/environment to verify performance (need to be contending with usual profile of reads/writes against db), i'm loathe to check anything in, however, as a starting point the basic sql can be simplified somewhat ( + extra options at start of method to be included as WHERE conditions) to something like: ``` SELECT DISTINCT files.*, users.display_name, GROUP_CONCAT(tag SEPARATOR ' ') FROM gpx_files AS files INNER JOIN users ON files.user_id = users.id LEFT JOIN gpx_file_tags AS tags ON files.id = tags.gpx_id WHERE private = 0 AND users.display_name != '' AND visible = 1 GROUP BY files.id ORDER BY timestamp DESC LIMIT 20; ``` from mysql EXPLAIN it would help to have a couple of new indexes - though unsure of amount of impact: ``` CREATE INDEX users_display_name_idx ON users(display_name); CREATE INDEX gpx_files_visible_public_idx ON gpx_files(visible, public); ``` and this one from API 0.4: ``` CREATE INDEX gpx_file_tags_gpxid_idx ON gpx_file_tags(gpx_id); ```
1.0
API 0.3 traces display has slow query - **[Submitted to the original trac issue database at 6.42pm, Friday, 23rd March 2007]** found at source:/www.openstreetmap.org/ruby/api/osm/dao.rb@R1589#L431 seems hard to imagine that cross-reffing 20000 trace files and 5000 users could become a major drain on db, but that is apparently the situation. whatever the exact situation, the sql in here looks pretty grim. anyways, in the course of putting the traces stuff into 0.4 API i needed to check this code, so i thought i'd record results: without the test data/environment to verify performance (need to be contending with usual profile of reads/writes against db), i'm loathe to check anything in, however, as a starting point the basic sql can be simplified somewhat ( + extra options at start of method to be included as WHERE conditions) to something like: ``` SELECT DISTINCT files.*, users.display_name, GROUP_CONCAT(tag SEPARATOR ' ') FROM gpx_files AS files INNER JOIN users ON files.user_id = users.id LEFT JOIN gpx_file_tags AS tags ON files.id = tags.gpx_id WHERE private = 0 AND users.display_name != '' AND visible = 1 GROUP BY files.id ORDER BY timestamp DESC LIMIT 20; ``` from mysql EXPLAIN it would help to have a couple of new indexes - though unsure of amount of impact: ``` CREATE INDEX users_display_name_idx ON users(display_name); CREATE INDEX gpx_files_visible_public_idx ON gpx_files(visible, public); ``` and this one from API 0.4: ``` CREATE INDEX gpx_file_tags_gpxid_idx ON gpx_file_tags(gpx_id); ```
defect
api traces display has slow query found at source seems hard to imagine that cross reffing trace files and users could become a major drain on db but that is apparently the situation whatever the exact situation the sql in here looks pretty grim anyways in the course of putting the traces stuff into api i needed to check this code so i thought i d record results without the test data environment to verify performance need to be contending with usual profile of reads writes against db i m loathe to check anything in however as a starting point the basic sql can be simplified somewhat extra options at start of method to be included as where conditions to something like select distinct files users display name group concat tag separator from gpx files as files inner join users on files user id users id left join gpx file tags as tags on files id tags gpx id where private and users display name and visible group by files id order by timestamp desc limit from mysql explain it would help to have a couple of new indexes though unsure of amount of impact create index users display name idx on users display name create index gpx files visible public idx on gpx files visible public and this one from api create index gpx file tags gpxid idx on gpx file tags gpx id
1
185,223
15,015,086,337
IssuesEvent
2021-02-01 07:46:26
NotOnlyStudents/Documentazione
https://api.github.com/repos/NotOnlyStudents/Documentazione
opened
Cambiare categoria di RQO10
adr bug documentation
Tullio: "RQO10 è un requisito di qualità poiché non insiste sul processo, ma sul prodotto." Io e Filippo crediamo che vada messo in quelli di vincolo, ma conviene aspettare ciò che ci dice tullione che è una frase un po' ambigua
1.0
Cambiare categoria di RQO10 - Tullio: "RQO10 è un requisito di qualità poiché non insiste sul processo, ma sul prodotto." Io e Filippo crediamo che vada messo in quelli di vincolo, ma conviene aspettare ciò che ci dice tullione che è una frase un po' ambigua
non_defect
cambiare categoria di tullio è un requisito di qualità poiché non insiste sul processo ma sul prodotto io e filippo crediamo che vada messo in quelli di vincolo ma conviene aspettare ciò che ci dice tullione che è una frase un po ambigua
0