Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
39,673
9,608,808,635
IssuesEvent
2019-05-12 10:00:29
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
[Android] Collapase menu button is replaced with Back button
C: Android - Manager E: 1 day P: Minor R: fixed T: Defect
**Describe the bug** Top left menu button is replaced with a back button that is from Event log Is this intentional? **Steps To Reproduce** 1. Use the beta 2. Immediately see the behaviour **Expected behavior** "Hamburger" or collapse menu button should be still there... **Screenshots** The beta 7.14 ![img](https://user-images.githubusercontent.com/16503773/50486561-dc0fae80-09fa-11e9-8d20-b571ddf9de39.png) 7.4.53 ![screenshot_20181228-100737](https://user-images.githubusercontent.com/12658589/50500163-60f5d900-0a8a-11e9-96a2-dc9b81a41339.png) ![screenshot_20181228-100748](https://user-images.githubusercontent.com/12658589/50500138-3e63c000-0a8a-11e9-8d8b-569fd463cee5.png) **System Information (please complete the following information):** - OS: Android 8.1
1.0
[Android] Collapase menu button is replaced with Back button - **Describe the bug** Top left menu button is replaced with a back button that is from Event log Is this intentional? **Steps To Reproduce** 1. Use the beta 2. Immediately see the behaviour **Expected behavior** "Hamburger" or collapse menu button should be still there... **Screenshots** The beta 7.14 ![img](https://user-images.githubusercontent.com/16503773/50486561-dc0fae80-09fa-11e9-8d20-b571ddf9de39.png) 7.4.53 ![screenshot_20181228-100737](https://user-images.githubusercontent.com/12658589/50500163-60f5d900-0a8a-11e9-96a2-dc9b81a41339.png) ![screenshot_20181228-100748](https://user-images.githubusercontent.com/12658589/50500138-3e63c000-0a8a-11e9-8d8b-569fd463cee5.png) **System Information (please complete the following information):** - OS: Android 8.1
defect
collapase menu button is replaced with back button describe the bug top left menu button is replaced with a back button that is from event log is this intentional steps to reproduce use the beta immediately see the behaviour expected behavior hamburger or collapse menu button should be still there screenshots the beta system information please complete the following information os android
1
4,519
5,133,146,702
IssuesEvent
2017-01-11 02:01:56
archco/cosmos-css
https://api.github.com/repos/archco/cosmos-css
opened
Variables 정리
css enhancement infrastructure
#78 처리중에 알게된 sass의 색 변수에 대한 정리 http://sass-lang.com/styleguide/color 를 참고삼아 Theme와 관련된 변수들을 정리하자. ### Theme Colors List * primary-color * secondary-color * text-color * text-color-invert * text-color-weak * text-color-strong * body-bg-color * body-bg-color-shade * body-bg-color-invert * shadow-color * link-color * link-color-hover * link-color-visited * border-color
1.0
Variables 정리 - #78 처리중에 알게된 sass의 색 변수에 대한 정리 http://sass-lang.com/styleguide/color 를 참고삼아 Theme와 관련된 변수들을 정리하자. ### Theme Colors List * primary-color * secondary-color * text-color * text-color-invert * text-color-weak * text-color-strong * body-bg-color * body-bg-color-shade * body-bg-color-invert * shadow-color * link-color * link-color-hover * link-color-visited * border-color
non_defect
variables 정리 처리중에 알게된 sass의 색 변수에 대한 정리 를 참고삼아 theme와 관련된 변수들을 정리하자 theme colors list primary color secondary color text color text color invert text color weak text color strong body bg color body bg color shade body bg color invert shadow color link color link color hover link color visited border color
0
217,324
24,329,026,992
IssuesEvent
2022-09-30 17:29:58
hapifhir/hapi-fhir
https://api.github.com/repos/hapifhir/hapi-fhir
closed
CVE-2022-40154 (High) detected in woodstox-core-6.3.1.jar - autoclosed
security vulnerability
## CVE-2022-40154 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.3.1.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /hapi-fhir-structures-dstu3/pom.xml</p> <p>Path to vulnerable library: /repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar</p> <p> Dependency Hierarchy: - :x: **woodstox-core-6.3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430">ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to serialise XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40154>CVE-2022-40154</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-40154 (High) detected in woodstox-core-6.3.1.jar - autoclosed - ## CVE-2022-40154 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.3.1.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /hapi-fhir-structures-dstu3/pom.xml</p> <p>Path to vulnerable library: /repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar</p> <p> Dependency Hierarchy: - :x: **woodstox-core-6.3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430">ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to serialise XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40154>CVE-2022-40154</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in woodstox core jar autoclosed cve high severity vulnerability vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file hapi fhir structures pom xml path to vulnerable library repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy x woodstox core jar vulnerable library found in head commit a href found in base branch master vulnerability details those using xstream to serialise xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stack overflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
16,025
2,870,252,244
IssuesEvent
2015-06-07 00:37:28
pdelia/away3d
https://api.github.com/repos/pdelia/away3d
closed
Two view.render() calls required to update geometry correctly
auto-migrated Priority-Medium Type-Defect
#79 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:38Z ``` What steps will reproduce the problem? 1. Compile and run the attached FlashDevelop project for FP10, Away3D trunk (issue occurs in rev 2028 at least) 2. Press '>' to use two render calls. Press '<' to use a single render call. Press Space to swap cameras What is the expected output? What do you see instead? The 3D model consists of a camera and a sphere in an ObjectContainer3D. The camera is pointing directly at the sphere and so the sphere should be rendered in the centre of the frame. When the ObjectContainer3D is rotated and the view rendered the sphere appears offset from its correct location. This problem can be worked around by calling view.render() twice in succession. ``` Original issue reported on code.google.com by `brucewei...@googlemail.com` on 16 Nov 2009 at 6:43 Attachments: * [ContainerProblem.zip](https://storage.googleapis.com/google-code-attachments/away3d/issue-79/comment-0/ContainerProblem.zip)
1.0
Two view.render() calls required to update geometry correctly - #79 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:38Z ``` What steps will reproduce the problem? 1. Compile and run the attached FlashDevelop project for FP10, Away3D trunk (issue occurs in rev 2028 at least) 2. Press '>' to use two render calls. Press '<' to use a single render call. Press Space to swap cameras What is the expected output? What do you see instead? The 3D model consists of a camera and a sphere in an ObjectContainer3D. The camera is pointing directly at the sphere and so the sphere should be rendered in the centre of the frame. When the ObjectContainer3D is rotated and the view rendered the sphere appears offset from its correct location. This problem can be worked around by calling view.render() twice in succession. ``` Original issue reported on code.google.com by `brucewei...@googlemail.com` on 16 Nov 2009 at 6:43 Attachments: * [ContainerProblem.zip](https://storage.googleapis.com/google-code-attachments/away3d/issue-79/comment-0/ContainerProblem.zip)
defect
two view render calls required to update geometry correctly issue by googlecodeexporter created on what steps will reproduce the problem compile and run the attached flashdevelop project for trunk issue occurs in rev at least press to use two render calls press to use a single render call press space to swap cameras what is the expected output what do you see instead the model consists of a camera and a sphere in an the camera is pointing directly at the sphere and so the sphere should be rendered in the centre of the frame when the is rotated and the view rendered the sphere appears offset from its correct location this problem can be worked around by calling view render twice in succession original issue reported on code google com by brucewei googlemail com on nov at attachments
1
23,390
3,807,888,227
IssuesEvent
2016-03-25 11:39:20
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
pdns crashes when database insert fails to postgres database (lowercase constraint)
auth defect
In the current pdns setup, when the AXFR transfer contains names which has uppercase letters the following happens: ``` Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Transaction started for 'nnn.fi.' Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Unable to feed record during incoming AXFR of 'nnn.fi.': GSQLBackend unable to feed record: Fatal error during query: insert into records (content,ttl,prio,type,domain_id,disabled,name,ordername,auth,change_date) values ($1,$2,$3,$4,$5,$6,$7,$8,$9,null): ERROR: new row for relation "records" violates check constraint "c_lowercase_name" Mar 16 04:58:37 ns3 pdns_server: DETAIL: Failing row contains (6802, 5, 0A.client.nnn.fi, A, 192.168.0.10, 300, 0, null, f, null, t). Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Aborting possible open transaction for domain 'nnn.fi.' AXFR Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Got a signal 11, attempting to print trace: Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0x18195b) [0x7fde2958495b] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(+0x35670) [0x7fde27948670] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/pgsql-9.5/lib/libpq.so.5(PQclear+0x69) [0x7fde272c0f49] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN15SPgSQLStatementD0Ev+0x5d) [0x7fde274f4f5d] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN11GSQLBackend14freeStatementsEv+0x199) [0x7fde294967d9] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN13gPgSQLBackendD0Ev+0x34) [0x7fde274f1d24] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackend7cleanupEv+0x184) [0x7fde295d9f84] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackendD2Ev+0x2b) [0x7fde295da03b] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass4suckERK7DNSNameRKSs+0x5fc) [0x7fde295b211c] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass19retrievalLoopThreadEv+0x136) [0x7fde294aded6] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0xaea69) [0x7fde294b1a69] Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libpthread.so.0(+0x7dc5) [0x7fde27cdbdc5] Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(clone+0x6d) [0x7fde27a0928d] ```
1.0
pdns crashes when database insert fails to postgres database (lowercase constraint) - In the current pdns setup, when the AXFR transfer contains names which has uppercase letters the following happens: ``` Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Transaction started for 'nnn.fi.' Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Unable to feed record during incoming AXFR of 'nnn.fi.': GSQLBackend unable to feed record: Fatal error during query: insert into records (content,ttl,prio,type,domain_id,disabled,name,ordername,auth,change_date) values ($1,$2,$3,$4,$5,$6,$7,$8,$9,null): ERROR: new row for relation "records" violates check constraint "c_lowercase_name" Mar 16 04:58:37 ns3 pdns_server: DETAIL: Failing row contains (6802, 5, 0A.client.nnn.fi, A, 192.168.0.10, 300, 0, null, f, null, t). Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Aborting possible open transaction for domain 'nnn.fi.' AXFR Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Got a signal 11, attempting to print trace: Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0x18195b) [0x7fde2958495b] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(+0x35670) [0x7fde27948670] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/pgsql-9.5/lib/libpq.so.5(PQclear+0x69) [0x7fde272c0f49] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN15SPgSQLStatementD0Ev+0x5d) [0x7fde274f4f5d] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN11GSQLBackend14freeStatementsEv+0x199) [0x7fde294967d9] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN13gPgSQLBackendD0Ev+0x34) [0x7fde274f1d24] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackend7cleanupEv+0x184) [0x7fde295d9f84] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackendD2Ev+0x2b) [0x7fde295da03b] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass4suckERK7DNSNameRKSs+0x5fc) [0x7fde295b211c] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass19retrievalLoopThreadEv+0x136) [0x7fde294aded6] Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0xaea69) [0x7fde294b1a69] Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libpthread.so.0(+0x7dc5) [0x7fde27cdbdc5] Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(clone+0x6d) [0x7fde27a0928d] ```
defect
pdns crashes when database insert fails to postgres database lowercase constraint in the current pdns setup when the axfr transfer contains names which has uppercase letters the following happens mar pdns server mar transaction started for nnn fi mar pdns server mar unable to feed record during incoming axfr of nnn fi gsqlbackend unable to feed record fatal error during query insert into records content ttl prio type domain id disabled name ordername auth change date values null error new row for relation records violates check constraint c lowercase name mar pdns server detail failing row contains client nnn fi a null f null t mar pdns server mar aborting possible open transaction for domain nnn fi axfr mar pdns server mar got a signal attempting to print trace mar pdns server mar usr sbin pdns server mar pdns server mar libc so mar pdns server mar usr pgsql lib libpq so pqclear mar pdns server mar usr pdns libgpgsqlbackend so mar pdns server mar usr sbin pdns server mar pdns server mar usr pdns libgpgsqlbackend so mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar libpthread so mar pdns server mar libc so clone
1
7,356
2,610,364,949
IssuesEvent
2015-02-26 19:57:56
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Config save failing
auto-migrated Priority-Medium tumblr Type-Defect
``` What's the problem? Wordpress.com and tumblr.com config fail to save. Login info has been confirmed. What browser are you using? Chrome + Windows What version of ScribeFire are you running? ScribeFire 4.1 ``` ----- Original issue reported on code.google.com by `alexande...@sequoiavision.com` on 21 Dec 2012 at 3:12
1.0
Config save failing - ``` What's the problem? Wordpress.com and tumblr.com config fail to save. Login info has been confirmed. What browser are you using? Chrome + Windows What version of ScribeFire are you running? ScribeFire 4.1 ``` ----- Original issue reported on code.google.com by `alexande...@sequoiavision.com` on 21 Dec 2012 at 3:12
defect
config save failing what s the problem wordpress com and tumblr com config fail to save login info has been confirmed what browser are you using chrome windows what version of scribefire are you running scribefire original issue reported on code google com by alexande sequoiavision com on dec at
1
30,666
6,219,850,217
IssuesEvent
2017-07-09 17:23:12
pymc-devs/pymc3
https://api.github.com/repos/pymc-devs/pymc3
closed
sample_smc: AttributeError: 'NoneType' object has no attribute 'n_steps'
defects
``` with pm.Model(): pm.NormalMixture('m', mu=np.array([0., 1.]), w=np.array([.6, .4]), sd=np.array([.2, .2])) step = pm.smc.SMC(n_chains=500) trace = pm.smc.sample_smc(n_steps=1000, step=step, homepath='/tmp') ``` `traceplot(trace)` gives: ![image](https://user-images.githubusercontent.com/674200/27769932-025816ac-5f36-11e7-9fca-a9499857b86d.png) `sns.distplot(trace.get_values('m'));` gives: ![image](https://user-images.githubusercontent.com/674200/27769934-0bf06408-5f36-11e7-9fd9-875294f9cf1a.png)
1.0
sample_smc: AttributeError: 'NoneType' object has no attribute 'n_steps' - ``` with pm.Model(): pm.NormalMixture('m', mu=np.array([0., 1.]), w=np.array([.6, .4]), sd=np.array([.2, .2])) step = pm.smc.SMC(n_chains=500) trace = pm.smc.sample_smc(n_steps=1000, step=step, homepath='/tmp') ``` `traceplot(trace)` gives: ![image](https://user-images.githubusercontent.com/674200/27769932-025816ac-5f36-11e7-9fca-a9499857b86d.png) `sns.distplot(trace.get_values('m'));` gives: ![image](https://user-images.githubusercontent.com/674200/27769934-0bf06408-5f36-11e7-9fd9-875294f9cf1a.png)
defect
sample smc attributeerror nonetype object has no attribute n steps with pm model pm normalmixture m mu np array w np array sd np array step pm smc smc n chains trace pm smc sample smc n steps step step homepath tmp traceplot trace gives sns distplot trace get values m gives
1
81,309
30,791,832,407
IssuesEvent
2023-07-31 16:44:09
vector-im/element-desktop
https://api.github.com/repos/vector-im/element-desktop
opened
MacOS: Unable to access microphone when try to call
T-Defect
### Steps to reproduce 1. Upgrade the formerly working client to current version (up from 1.11.13) on a MacOS Ventura machine. 2. As soon as you start a call you receive the failure message. If you get a call there's not even the button to take it. 3. Downgrade to the last working version 1.11.13 restores the ability to call and take calls. ### Outcome #### What did you expect? I expected the Calls feature to keep working when upgrading to a more current version of Element. #### What happened instead? When I try to start a voice/video call, I get the following error and can not start the call: ```` Unable to access microphone Call failed because microphone could not be accessed. Check that a microphone is plugged in and set up correctly. ```` ### Operating system MacOS ### Application version Element higher than 1.11.13 ### How did you install the app? https://element.io ### Homeserver docker image matrixdotorg/synapse:latest ### Will you send logs? Yes
1.0
MacOS: Unable to access microphone when try to call - ### Steps to reproduce 1. Upgrade the formerly working client to current version (up from 1.11.13) on a MacOS Ventura machine. 2. As soon as you start a call you receive the failure message. If you get a call there's not even the button to take it. 3. Downgrade to the last working version 1.11.13 restores the ability to call and take calls. ### Outcome #### What did you expect? I expected the Calls feature to keep working when upgrading to a more current version of Element. #### What happened instead? When I try to start a voice/video call, I get the following error and can not start the call: ```` Unable to access microphone Call failed because microphone could not be accessed. Check that a microphone is plugged in and set up correctly. ```` ### Operating system MacOS ### Application version Element higher than 1.11.13 ### How did you install the app? https://element.io ### Homeserver docker image matrixdotorg/synapse:latest ### Will you send logs? Yes
defect
macos unable to access microphone when try to call steps to reproduce upgrade the formerly working client to current version up from on a macos ventura machine as soon as you start a call you receive the failure message if you get a call there s not even the button to take it downgrade to the last working version restores the ability to call and take calls outcome what did you expect i expected the calls feature to keep working when upgrading to a more current version of element what happened instead when i try to start a voice video call i get the following error and can not start the call unable to access microphone call failed because microphone could not be accessed check that a microphone is plugged in and set up correctly operating system macos application version element higher than how did you install the app homeserver docker image matrixdotorg synapse latest will you send logs yes
1
433,111
30,313,058,391
IssuesEvent
2023-07-10 13:58:32
SciTools/iris
https://api.github.com/repos/SciTools/iris
opened
Change the default version of the documentation to stable rather than latest on ReadTheDocs
Type: Documentation
## 📚 Documentation <!-- See https://scitools-iris.readthedocs.io/en/latest/ --> <!-- Describe the issue or provide a suggestion for improving the Iris documentation --> Currently the default version for the documentation on readthedocs is latest, which is the documentation for the next, unreleased version of Iris. This should be changed to the stable, which is the most recent release. Because the unreleased documentation is the default, it is what is directed to when you visit the top of the domain ([scitools-iris.readthedocs.io](https://scitools-iris.readthedocs.io/)), as well as affecting the [canonical URL](https://docs.readthedocs.io/en/latest/canonical-urls.html#how-read-the-docs-generates-canonical-urls), which makes search engines point to the unreleased documentation in over the stable docs. The current behaviour is both confusing, and [not recommended](https://docs.readthedocs.io/en/latest/guides/best-practice/links.html#questionable-practice). Fortunately it can be changed fairly easily in the ReadTheDocs web interface. The [setting](https://docs.readthedocs.io/en/latest/config-file/v2.html#migrating-from-the-web-interface) is `Admin` > `Advanced settings` > `Default version`.
1.0
Change the default version of the documentation to stable rather than latest on ReadTheDocs - ## 📚 Documentation <!-- See https://scitools-iris.readthedocs.io/en/latest/ --> <!-- Describe the issue or provide a suggestion for improving the Iris documentation --> Currently the default version for the documentation on readthedocs is latest, which is the documentation for the next, unreleased version of Iris. This should be changed to the stable, which is the most recent release. Because the unreleased documentation is the default, it is what is directed to when you visit the top of the domain ([scitools-iris.readthedocs.io](https://scitools-iris.readthedocs.io/)), as well as affecting the [canonical URL](https://docs.readthedocs.io/en/latest/canonical-urls.html#how-read-the-docs-generates-canonical-urls), which makes search engines point to the unreleased documentation in over the stable docs. The current behaviour is both confusing, and [not recommended](https://docs.readthedocs.io/en/latest/guides/best-practice/links.html#questionable-practice). Fortunately it can be changed fairly easily in the ReadTheDocs web interface. The [setting](https://docs.readthedocs.io/en/latest/config-file/v2.html#migrating-from-the-web-interface) is `Admin` > `Advanced settings` > `Default version`.
non_defect
change the default version of the documentation to stable rather than latest on readthedocs 📚 documentation currently the default version for the documentation on readthedocs is latest which is the documentation for the next unreleased version of iris this should be changed to the stable which is the most recent release because the unreleased documentation is the default it is what is directed to when you visit the top of the domain as well as affecting the which makes search engines point to the unreleased documentation in over the stable docs the current behaviour is both confusing and fortunately it can be changed fairly easily in the readthedocs web interface the is admin advanced settings default version
0
46,237
9,915,049,486
IssuesEvent
2019-06-28 15:48:36
Badgerati/Pode
https://api.github.com/repos/Badgerati/Pode
opened
Split the flash function into multiple PowerShell functions
internal-code :hammer:
The `flash` function's actions should be split out into multiple PowerShell functions. Each of these functions should be exported in `pode.psd1`. | Action | New Function | Returns | | ------ | ------------ | ------- | | `Add` | `Add-PodeFlashMessage` | `void` | | `Clear` | `Clear-PodeFlashMessages` | `void` | | `Get` | `Get-PodeFlashMessage` | `array` | | `Keys` | `Get-PodeFlashMessageKeys` | `array` | | `Remove` | `Remove-PodeFlashMessage` | `bool` | Possible new functions: * `Test-PodeFlashMessageKey` (test if a key is present)
1.0
Split the flash function into multiple PowerShell functions - The `flash` function's actions should be split out into multiple PowerShell functions. Each of these functions should be exported in `pode.psd1`. | Action | New Function | Returns | | ------ | ------------ | ------- | | `Add` | `Add-PodeFlashMessage` | `void` | | `Clear` | `Clear-PodeFlashMessages` | `void` | | `Get` | `Get-PodeFlashMessage` | `array` | | `Keys` | `Get-PodeFlashMessageKeys` | `array` | | `Remove` | `Remove-PodeFlashMessage` | `bool` | Possible new functions: * `Test-PodeFlashMessageKey` (test if a key is present)
non_defect
split the flash function into multiple powershell functions the flash function s actions should be split out into multiple powershell functions each of these functions should be exported in pode action new function returns add add podeflashmessage void clear clear podeflashmessages void get get podeflashmessage array keys get podeflashmessagekeys array remove remove podeflashmessage bool possible new functions test podeflashmessagekey test if a key is present
0
139,582
12,875,744,979
IssuesEvent
2020-07-11 00:31:09
dotnet/diagnostics
https://api.github.com/repos/dotnet/diagnostics
opened
Add aks.ms link to diagnostics FAQ in SOS error messages
documentation dotnet-sos
And document in the FAQ the reasons for this message: ``` Failed to find runtime module (libcoreclr.so), 0x80004005 Extension commands need it in order to have something to do. ClrStack failed ``` Some of the reasons for this message are: 1) the process or core dump hasn't loaded the .NET Core runtime yet. 2) lldb on a coredump wasn't started with the host program . `target modules list` doesn't display libcoreclr.so. 3) lldb on a coredump needs the correct version libcoreclr.so on disk. `target modules list` displays libcoreclr.so but SOS commands display this message. Use `dotnet-symbol --modules <coredump>` to download libcoreclr.so from the coredump.
1.0
Add aks.ms link to diagnostics FAQ in SOS error messages - And document in the FAQ the reasons for this message: ``` Failed to find runtime module (libcoreclr.so), 0x80004005 Extension commands need it in order to have something to do. ClrStack failed ``` Some of the reasons for this message are: 1) the process or core dump hasn't loaded the .NET Core runtime yet. 2) lldb on a coredump wasn't started with the host program . `target modules list` doesn't display libcoreclr.so. 3) lldb on a coredump needs the correct version libcoreclr.so on disk. `target modules list` displays libcoreclr.so but SOS commands display this message. Use `dotnet-symbol --modules <coredump>` to download libcoreclr.so from the coredump.
non_defect
add aks ms link to diagnostics faq in sos error messages and document in the faq the reasons for this message failed to find runtime module libcoreclr so extension commands need it in order to have something to do clrstack failed some of the reasons for this message are the process or core dump hasn t loaded the net core runtime yet lldb on a coredump wasn t started with the host program target modules list doesn t display libcoreclr so lldb on a coredump needs the correct version libcoreclr so on disk target modules list displays libcoreclr so but sos commands display this message use dotnet symbol modules to download libcoreclr so from the coredump
0
24,168
3,922,117,522
IssuesEvent
2016-04-22 03:49:32
furushchev/alchemy-2
https://api.github.com/repos/furushchev/alchemy-2
closed
Weight 10 not working
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? bin/infer -i test.mln -e empty.db -r results.txt -q good What is the expected output? What do you see instead? expected (actually produced for weights 9.999999 and 10.000001) good(A) 4.9995e-05 instead good(A) 0.99995 What version of the product are you using? On what operating system? Latest release (Jan 2013) on debian wheezy Please provide any additional information below. All other weights work just fine, but for some reason 10 doesn't work. ``` Original issue reported on code.google.com by `ads...@gmail.com` on 19 Jan 2014 at 9:11 Attachments: * [test.mln](https://storage.googleapis.com/google-code-attachments/alchemy-2/issue-6/comment-0/test.mln)
1.0
Weight 10 not working - ``` What steps will reproduce the problem? bin/infer -i test.mln -e empty.db -r results.txt -q good What is the expected output? What do you see instead? expected (actually produced for weights 9.999999 and 10.000001) good(A) 4.9995e-05 instead good(A) 0.99995 What version of the product are you using? On what operating system? Latest release (Jan 2013) on debian wheezy Please provide any additional information below. All other weights work just fine, but for some reason 10 doesn't work. ``` Original issue reported on code.google.com by `ads...@gmail.com` on 19 Jan 2014 at 9:11 Attachments: * [test.mln](https://storage.googleapis.com/google-code-attachments/alchemy-2/issue-6/comment-0/test.mln)
defect
weight not working what steps will reproduce the problem bin infer i test mln e empty db r results txt q good what is the expected output what do you see instead expected actually produced for weights and good a instead good a what version of the product are you using on what operating system latest release jan on debian wheezy please provide any additional information below all other weights work just fine but for some reason doesn t work original issue reported on code google com by ads gmail com on jan at attachments
1
68,448
21,664,594,704
IssuesEvent
2022-05-07 01:53:13
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
New reactions picker sometimes goes off the top of the screen
T-Defect A-Reactions
I'll do a screenshot when I see it again - from memory I was reacting to a messages just about halfway up the visible timeline. Firefox osx
1.0
New reactions picker sometimes goes off the top of the screen - I'll do a screenshot when I see it again - from memory I was reacting to a messages just about halfway up the visible timeline. Firefox osx
defect
new reactions picker sometimes goes off the top of the screen i ll do a screenshot when i see it again from memory i was reacting to a messages just about halfway up the visible timeline firefox osx
1
385,369
11,419,709,401
IssuesEvent
2020-02-03 08:37:23
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Exception handling issue in CacheCleanupTask
Priority/Highest Severity/Critical Type/Bug
We have the following code block in [1] to clear up caches that are expired or overflowed. There are multiple caches in the "caches" list and they are cleaned by invoking the `runCacheExpiry()` of each of them going through a loop. ``` try { ... for (CacheImpl cache : caches) { cache.runCacheExpiry(); } } } catch (IllegalStateException e) { log.debug("Error occurred while running CacheCleanupTask", e); ``` If `cache.runCacheExpiry()` throws an exception, the whole cleanup task will break and skips clearing the rest of the caches after this exception. Therefore, it possible that some caches not cleaned up for a longer time and get overflowed. Cache eviction also happening in the same method `cache.runCacheExpiry()`. Therefore, those non-cleaned caches do not accept new cache entries and different flows start to fail. ***Suggestion*** Catch IllegalStateException inside the loop itself and avoid failing the whole CacheCleanupTask [1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/javax.cache/src/main/java/org/wso2/carbon/caching/impl/CacheCleanupTask.java#L47
1.0
Exception handling issue in CacheCleanupTask - We have the following code block in [1] to clear up caches that are expired or overflowed. There are multiple caches in the "caches" list and they are cleaned by invoking the `runCacheExpiry()` of each of them going through a loop. ``` try { ... for (CacheImpl cache : caches) { cache.runCacheExpiry(); } } } catch (IllegalStateException e) { log.debug("Error occurred while running CacheCleanupTask", e); ``` If `cache.runCacheExpiry()` throws an exception, the whole cleanup task will break and skips clearing the rest of the caches after this exception. Therefore, it possible that some caches not cleaned up for a longer time and get overflowed. Cache eviction also happening in the same method `cache.runCacheExpiry()`. Therefore, those non-cleaned caches do not accept new cache entries and different flows start to fail. ***Suggestion*** Catch IllegalStateException inside the loop itself and avoid failing the whole CacheCleanupTask [1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/javax.cache/src/main/java/org/wso2/carbon/caching/impl/CacheCleanupTask.java#L47
non_defect
exception handling issue in cachecleanuptask we have the following code block in to clear up caches that are expired or overflowed there are multiple caches in the caches list and they are cleaned by invoking the runcacheexpiry of each of them going through a loop try for cacheimpl cache caches cache runcacheexpiry catch illegalstateexception e log debug error occurred while running cachecleanuptask e if cache runcacheexpiry throws an exception the whole cleanup task will break and skips clearing the rest of the caches after this exception therefore it possible that some caches not cleaned up for a longer time and get overflowed cache eviction also happening in the same method cache runcacheexpiry therefore those non cleaned caches do not accept new cache entries and different flows start to fail suggestion catch illegalstateexception inside the loop itself and avoid failing the whole cachecleanuptask
0
234,664
7,724,418,864
IssuesEvent
2018-05-24 15:02:13
test4gloirin/m
https://api.github.com/repos/test4gloirin/m
closed
0000074: replace getXXX/countXXX functions
Addressbook Feature Request low priority
**Reported by pschuele on 14 Jul 2008 15:17** All getXXX/countXXX in json/controller/backend classes should be replaced by one get and one count function (see crm and task applications). javascript frontend has to be updated.
1.0
0000074: replace getXXX/countXXX functions - **Reported by pschuele on 14 Jul 2008 15:17** All getXXX/countXXX in json/controller/backend classes should be replaced by one get and one count function (see crm and task applications). javascript frontend has to be updated.
non_defect
replace getxxx countxxx functions reported by pschuele on jul all getxxx countxxx in json controller backend classes should be replaced by one get and one count function see crm and task applications javascript frontend has to be updated
0
33,009
6,994,840,804
IssuesEvent
2017-12-15 16:47:54
contao/installation-bundle
https://api.github.com/repos/contao/installation-bundle
closed
The MySQL version check cannot determine the current version
defect
It appears that the [check for the MySQL version](https://github.com/contao/installation-bundle/commit/0514c650bce809bc8081badcdb1cda291f31eecf) may not always correctly determine the current version: ![screenshot-2017-12-15 contao open source cms](https://user-images.githubusercontent.com/1218766/34021409-0399d2d6-e13a-11e7-9e83-6d7726d9a636.png) Tested with Contao 4.5.0-RC1 and `mysql Ver 14.14 Distrib 5.1.73, for redhat-linux-gnu (x86_64)`
1.0
The MySQL version check cannot determine the current version - It appears that the [check for the MySQL version](https://github.com/contao/installation-bundle/commit/0514c650bce809bc8081badcdb1cda291f31eecf) may not always correctly determine the current version: ![screenshot-2017-12-15 contao open source cms](https://user-images.githubusercontent.com/1218766/34021409-0399d2d6-e13a-11e7-9e83-6d7726d9a636.png) Tested with Contao 4.5.0-RC1 and `mysql Ver 14.14 Distrib 5.1.73, for redhat-linux-gnu (x86_64)`
defect
the mysql version check cannot determine the current version it appears that the may not always correctly determine the current version tested with contao and mysql ver distrib for redhat linux gnu
1
289,255
24,971,143,150
IssuesEvent
2022-11-02 01:21:27
CMU-313/fall-22-hw4-team-sweg
https://api.github.com/repos/CMU-313/fall-22-hw4-team-sweg
closed
Test API to delete a model
test
Currently, there are no tests that make sure the delete function works as expected and takes in an appropriate model_id. I will add tests to test_models.py that will assert that the delete works as expected.
1.0
Test API to delete a model - Currently, there are no tests that make sure the delete function works as expected and takes in an appropriate model_id. I will add tests to test_models.py that will assert that the delete works as expected.
non_defect
test api to delete a model currently there are no tests that make sure the delete function works as expected and takes in an appropriate model id i will add tests to test models py that will assert that the delete works as expected
0
30,734
6,261,182,665
IssuesEvent
2017-07-14 23:00:21
blong42/clearsilver
https://api.github.com/repos/blong42/clearsilver
closed
problem importing neo_cgi : missing crc32
auto-migrated Priority-Medium Type-Defect
``` $ cd /usr/local/lib/python2.4/site-packages/site-packages$ python -c "import neo_cgi" Traceback (most recent call last): File "<string>", line 1, in ? ImportError: ./neo_cgi.so: undefined symbol: crc32 I am sure you assumed it would come from zlib.h ... not on my system! I imagine this has to do with Config.CompressionEnabled. Linux 3.2.0-29-generic #46-Ubuntu SMP Fri Jul 27 17:03:23 UTC 2012 x86_64 x86_64 x86_64 GNU/Linux ``` Original issue reported on code.google.com by `brianh...@gmail.com` on 16 Aug 2012 at 8:47
1.0
problem importing neo_cgi : missing crc32 - ``` $ cd /usr/local/lib/python2.4/site-packages/site-packages$ python -c "import neo_cgi" Traceback (most recent call last): File "<string>", line 1, in ? ImportError: ./neo_cgi.so: undefined symbol: crc32 I am sure you assumed it would come from zlib.h ... not on my system! I imagine this has to do with Config.CompressionEnabled. Linux 3.2.0-29-generic #46-Ubuntu SMP Fri Jul 27 17:03:23 UTC 2012 x86_64 x86_64 x86_64 GNU/Linux ``` Original issue reported on code.google.com by `brianh...@gmail.com` on 16 Aug 2012 at 8:47
defect
problem importing neo cgi missing cd usr local lib site packages site packages python c import neo cgi traceback most recent call last file line in importerror neo cgi so undefined symbol i am sure you assumed it would come from zlib h not on my system i imagine this has to do with config compressionenabled linux generic ubuntu smp fri jul utc gnu linux original issue reported on code google com by brianh gmail com on aug at
1
26,802
4,789,119,292
IssuesEvent
2016-10-30 22:17:26
belangeo/pyo
https://api.github.com/repos/belangeo/pyo
closed
Jack often zombifies on Ubuntu
auto-migrated Priority-Medium Type-Defect
``` Run any pyo script with Jack support. We often see "JACK error: zombified - calling shutdown handler". Sometimes it's OK, sometimes it's not. It's on Ubuntu Lucid, Natty, etc. ``` Original issue reported on code.google.com by `que...@gmail.com` on 21 Nov 2011 at 7:02
1.0
Jack often zombifies on Ubuntu - ``` Run any pyo script with Jack support. We often see "JACK error: zombified - calling shutdown handler". Sometimes it's OK, sometimes it's not. It's on Ubuntu Lucid, Natty, etc. ``` Original issue reported on code.google.com by `que...@gmail.com` on 21 Nov 2011 at 7:02
defect
jack often zombifies on ubuntu run any pyo script with jack support we often see jack error zombified calling shutdown handler sometimes it s ok sometimes it s not it s on ubuntu lucid natty etc original issue reported on code google com by que gmail com on nov at
1
407,074
11,906,191,221
IssuesEvent
2020-03-30 19:55:24
gsoft-inc/sg-orbit
https://api.github.com/repos/gsoft-inc/sg-orbit
closed
A select item with an icon doesn't follow text / icon ratio
bug components high priority theming
<!-- Please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo. If the bug prevent your team from releasing please add the tag "blocking" to your issue. --> ### Describe the bug ![image](https://user-images.githubusercontent.com/794579/77432169-16116100-6db4-11ea-999e-27c8bdba68e4.png) A select item icon size adapt to the select size but the text doesn't. Therefore, it breaks our standard text /icon ratio. Available options: - Don't adapt the icon size to the select size - Adapt the text size to the select size
1.0
A select item with an icon doesn't follow text / icon ratio - <!-- Please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo. If the bug prevent your team from releasing please add the tag "blocking" to your issue. --> ### Describe the bug ![image](https://user-images.githubusercontent.com/794579/77432169-16116100-6db4-11ea-999e-27c8bdba68e4.png) A select item icon size adapt to the select size but the text doesn't. Therefore, it breaks our standard text /icon ratio. Available options: - Don't adapt the icon size to the select size - Adapt the text size to the select size
non_defect
a select item with an icon doesn t follow text icon ratio please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo if the bug prevent your team from releasing please add the tag blocking to your issue describe the bug a select item icon size adapt to the select size but the text doesn t therefore it breaks our standard text icon ratio available options don t adapt the icon size to the select size adapt the text size to the select size
0
48,907
13,184,772,576
IssuesEvent
2020-08-12 20:03:58
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/423 , reported by claudio.kopper and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:33:36", "description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "_ts": "1351704816000000", "component": "combo reconstruction", "summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional", "priority": "normal", "keywords": "", "time": "2012-06-25T00:59:43", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/423 , reported by claudio.kopper and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:33:36", "description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "_ts": "1351704816000000", "component": "combo reconstruction", "summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional", "priority": "normal", "keywords": "", "time": "2012-06-25T00:59:43", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
sim services propagatorserviceutils propagate broken non functional trac migrated from reported by claudio kopper and owned by olivas json status closed changetime description i d like to commit the following attached patch with the following commit message n npropagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy n reporter claudio kopper cc resolution fixed ts component combo reconstruction summary sim services propagatorserviceutils propagate broken non functional priority normal keywords time milestone owner olivas type defect
1
375,963
11,136,509,261
IssuesEvent
2019-12-20 16:46:03
SynBioDex/SBOLExplorer
https://api.github.com/repos/SynBioDex/SBOLExplorer
closed
Create a separate SBOLExplorer log file
enhancement priority
This should be viewable on the SBOLExplorer admin page
1.0
Create a separate SBOLExplorer log file - This should be viewable on the SBOLExplorer admin page
non_defect
create a separate sbolexplorer log file this should be viewable on the sbolexplorer admin page
0
27,284
4,957,388,172
IssuesEvent
2016-12-02 04:09:36
google/gitiles
https://api.github.com/repos/google/gitiles
reopened
Add option for embedded (iframe) Gitiles views.
Priority-Medium Status-New Type-Defect
Originally reported on Google Code with ID 65 ``` It is useful for applications wishing to embed Gitiles data to have access to a trimmed-down form of Gitiles pages that omit specifically the Gitiles header and footer, focusing only on the requested content. This is useful for applications wishing to embed Gitiles data in (for example) an iframe. The ability to render a embedded view should be independent of the content of that view. In other words, the parameter should not use the "pretty" parameter. Instead, a "view" parameter is proposed. The default "view" value is the current view. However, alternative views can be supplied. In this case, the iframe-embeddable view would use "?view=embed". ``` Reported by `None` on 2014-09-17 16:23:32
1.0
Add option for embedded (iframe) Gitiles views. - Originally reported on Google Code with ID 65 ``` It is useful for applications wishing to embed Gitiles data to have access to a trimmed-down form of Gitiles pages that omit specifically the Gitiles header and footer, focusing only on the requested content. This is useful for applications wishing to embed Gitiles data in (for example) an iframe. The ability to render a embedded view should be independent of the content of that view. In other words, the parameter should not use the "pretty" parameter. Instead, a "view" parameter is proposed. The default "view" value is the current view. However, alternative views can be supplied. In this case, the iframe-embeddable view would use "?view=embed". ``` Reported by `None` on 2014-09-17 16:23:32
defect
add option for embedded iframe gitiles views originally reported on google code with id it is useful for applications wishing to embed gitiles data to have access to a trimmed down form of gitiles pages that omit specifically the gitiles header and footer focusing only on the requested content this is useful for applications wishing to embed gitiles data in for example an iframe the ability to render a embedded view should be independent of the content of that view in other words the parameter should not use the pretty parameter instead a view parameter is proposed the default view value is the current view however alternative views can be supplied in this case the iframe embeddable view would use view embed reported by none on
1
79
2,504,337,062
IssuesEvent
2015-01-10 02:59:32
eleybourn/Book-Catalogue
https://api.github.com/repos/eleybourn/Book-Catalogue
closed
Fit/finish - wrong heading for add book by title/author (search internet)
Defect In Beta
Adding a book by title and author (search internet) has the heading of ISBN search.
1.0
Fit/finish - wrong heading for add book by title/author (search internet) - Adding a book by title and author (search internet) has the heading of ISBN search.
defect
fit finish wrong heading for add book by title author search internet adding a book by title and author search internet has the heading of isbn search
1
64,104
18,213,973,278
IssuesEvent
2021-09-30 00:14:25
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
ExtraNodesetGenerator can’t find correct node using coord when mesh dimension lower than spatial dimension
T: defect P: normal
## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0. https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114 ## Steps to Reproduce <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
1.0
ExtraNodesetGenerator can’t find correct node using coord when mesh dimension lower than spatial dimension - ## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0. https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114 ## Steps to Reproduce <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
defect
extranodesetgenerator can’t find correct node using coord when mesh dimension lower than spatial dimension bug description when the mesh dimension is lower than spatial dimension say mesh dimension is spatial dimension is the code will only copy coordinate components of “c” into “p” which is not correct since c may not be steps to reproduce impact
1
24,370
3,968,652,748
IssuesEvent
2016-05-03 20:26:10
lester88a/snova
https://api.github.com/repos/lester88a/snova
closed
新版0.23.0 Linux(32位)编译后运行问题?
auto-migrated Priority-Medium Type-Defect
``` 环境: v0.23.0 ubuntu14.04 x32 1-因为没有看到32位版本,自己编译第一次没通过,删掉图标� ��以了OpenShift所有功能正常,Heroku封了不是本应用的问题。但 有个疑问:运行在后台通过log可以查看运行情况,但退出怎�� �办?我是直接klii了。是否我哪里没有搞对还是没办法自己结 束进程? 2-其实老版本java部署在一些支持java的空间很不错,比较稳定� ��宜被封。另外老版本的user_spac.json,user_hosts.conf文件是否不� ��作用了? 3-snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov a的原始作者吗,很崇拜! ``` Original issue reported on code.google.com by `bob.yang...@gmail.com` on 30 Oct 2014 at 2:48
1.0
新版0.23.0 Linux(32位)编译后运行问题? - ``` 环境: v0.23.0 ubuntu14.04 x32 1-因为没有看到32位版本,自己编译第一次没通过,删掉图标� ��以了OpenShift所有功能正常,Heroku封了不是本应用的问题。但 有个疑问:运行在后台通过log可以查看运行情况,但退出怎�� �办?我是直接klii了。是否我哪里没有搞对还是没办法自己结 束进程? 2-其实老版本java部署在一些支持java的空间很不错,比较稳定� ��宜被封。另外老版本的user_spac.json,user_hosts.conf文件是否不� ��作用了? 3-snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov a的原始作者吗,很崇拜! ``` Original issue reported on code.google.com by `bob.yang...@gmail.com` on 30 Oct 2014 at 2:48
defect
linux( )编译后运行问题? 环境: ,自己编译第一次没通过,删掉图标� ��以了openshift所有功能正常,heroku封了不是本应用的问题。但 有个疑问:运行在后台通过log可以查看运行情况,但退出怎�� �办?我是直接klii了。是否我哪里没有搞对还是没办法自己结 束进程? 其实老版本java部署在一些支持java的空间很不错,比较稳定� ��宜被封。另外老版本的user spac json,user hosts conf文件是否不� ��作用了? snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov a的原始作者吗,很崇拜! original issue reported on code google com by bob yang gmail com on oct at
1
73,258
24,527,706,886
IssuesEvent
2022-10-11 14:14:40
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
[Server] one_result_per_user_per_wu potentially broken
C: Server - Scheduler P: Minor R: worksforme T: Defect E: to be determined
**Describe the bug** The server configuration option `one_result_per_user_per_wu` seems to not function as expected. I've seen 3 separate BOINC servers that have this option enabled and some users get two tasks from the same workunit. If the `one_result_per_host_per_wu` option is not enabled, a single host can get both tasks from a workunit: ![Screenshot_from_2021-03-29_15-13-09](https://user-images.githubusercontent.com/13613437/113120235-5ed9d200-91df-11eb-91d1-db3f3bc258cf.png) I spoke with the SiDock@home admin and, at the time of that log entry, `one_result_per_user_per_wu` was enabled but `one_result_per_host_per_wu` was not. **Affected Projects** - Minecraft@Home - SiDock@home - my own private BOINC instance **Steps To Reproduce** 1. Enable the rule `one_result_per_user_per_wu` in config.xml. 2. Wait for some new tasks to be distributed. 3. Go to the ops panel, click 'Workunits' under 'Browse database', then paste this into the 'additional clauses' box: `id in (select r.workunitid from result r join result r2 on r.workunitid = r2.workunitid and r.hostid <> r2.hostid and r.userid = r2.userid)` 4. Alternatively, in a clean BOINC server test environment, simply attach one host and watch as it gets both replications. **Expected behavior** The query should return no results - hosts should not get two results from the same workunit. **Screenshots** my BOINC instance: ![firefox_2021-03-31_05-17-05_7547](https://user-images.githubusercontent.com/13613437/113121158-59c95280-91e0-11eb-819b-80596f1a4548.png) Minecraft@Home: ![1guy2results](https://user-images.githubusercontent.com/13613437/113121369-95641c80-91e0-11eb-9447-d51f267cfbb3.png) SiDock@home: ![Screenshot_from_2021-03-29_15-13-09](https://user-images.githubusercontent.com/13613437/113120235-5ed9d200-91df-11eb-91d1-db3f3bc258cf.png) **System Information** - OS: Debian 10 Buster (used by my private BOINC server) - BOINC Version: 1.2.0 (https://github.com/BOINC/boinc/tree/server_release/1/1.2) (version used by both my server and Minecraft@Home) - BOINC Version: 1.1.0 (version used by SiDock@home) **Additional Context** Minecraft@Home config.xml: https://github.com/minecrafthome/minecrafthome/blob/master/images/makeproject/project/config.xml#L35 Feel free to ask for additional clarification if I missed anything!
1.0
[Server] one_result_per_user_per_wu potentially broken - **Describe the bug** The server configuration option `one_result_per_user_per_wu` seems to not function as expected. I've seen 3 separate BOINC servers that have this option enabled and some users get two tasks from the same workunit. If the `one_result_per_host_per_wu` option is not enabled, a single host can get both tasks from a workunit: ![Screenshot_from_2021-03-29_15-13-09](https://user-images.githubusercontent.com/13613437/113120235-5ed9d200-91df-11eb-91d1-db3f3bc258cf.png) I spoke with the SiDock@home admin and, at the time of that log entry, `one_result_per_user_per_wu` was enabled but `one_result_per_host_per_wu` was not. **Affected Projects** - Minecraft@Home - SiDock@home - my own private BOINC instance **Steps To Reproduce** 1. Enable the rule `one_result_per_user_per_wu` in config.xml. 2. Wait for some new tasks to be distributed. 3. Go to the ops panel, click 'Workunits' under 'Browse database', then paste this into the 'additional clauses' box: `id in (select r.workunitid from result r join result r2 on r.workunitid = r2.workunitid and r.hostid <> r2.hostid and r.userid = r2.userid)` 4. Alternatively, in a clean BOINC server test environment, simply attach one host and watch as it gets both replications. **Expected behavior** The query should return no results - hosts should not get two results from the same workunit. **Screenshots** my BOINC instance: ![firefox_2021-03-31_05-17-05_7547](https://user-images.githubusercontent.com/13613437/113121158-59c95280-91e0-11eb-819b-80596f1a4548.png) Minecraft@Home: ![1guy2results](https://user-images.githubusercontent.com/13613437/113121369-95641c80-91e0-11eb-9447-d51f267cfbb3.png) SiDock@home: ![Screenshot_from_2021-03-29_15-13-09](https://user-images.githubusercontent.com/13613437/113120235-5ed9d200-91df-11eb-91d1-db3f3bc258cf.png) **System Information** - OS: Debian 10 Buster (used by my private BOINC server) - BOINC Version: 1.2.0 (https://github.com/BOINC/boinc/tree/server_release/1/1.2) (version used by both my server and Minecraft@Home) - BOINC Version: 1.1.0 (version used by SiDock@home) **Additional Context** Minecraft@Home config.xml: https://github.com/minecrafthome/minecrafthome/blob/master/images/makeproject/project/config.xml#L35 Feel free to ask for additional clarification if I missed anything!
defect
one result per user per wu potentially broken describe the bug the server configuration option one result per user per wu seems to not function as expected i ve seen separate boinc servers that have this option enabled and some users get two tasks from the same workunit if the one result per host per wu option is not enabled a single host can get both tasks from a workunit i spoke with the sidock home admin and at the time of that log entry one result per user per wu was enabled but one result per host per wu was not affected projects minecraft home sidock home my own private boinc instance steps to reproduce enable the rule one result per user per wu in config xml wait for some new tasks to be distributed go to the ops panel click workunits under browse database then paste this into the additional clauses box id in select r workunitid from result r join result on r workunitid workunitid and r hostid hostid and r userid userid alternatively in a clean boinc server test environment simply attach one host and watch as it gets both replications expected behavior the query should return no results hosts should not get two results from the same workunit screenshots my boinc instance minecraft home sidock home system information os debian buster used by my private boinc server boinc version version used by both my server and minecraft home boinc version version used by sidock home additional context minecraft home config xml feel free to ask for additional clarification if i missed anything
1
174,173
27,589,179,489
IssuesEvent
2023-03-08 22:30:06
NCIOCPL/cgov-digital-platform
https://api.github.com/repos/NCIOCPL/cgov-digital-platform
closed
update ncids version to use uswds normalize
Drupal - Redesign
## Issue description Bump package version of ncids to include uswds normalize file https://github.com/NCIOCPL/ncids/issues/738
1.0
update ncids version to use uswds normalize - ## Issue description Bump package version of ncids to include uswds normalize file https://github.com/NCIOCPL/ncids/issues/738
non_defect
update ncids version to use uswds normalize issue description bump package version of ncids to include uswds normalize file
0
365,279
10,780,354,687
IssuesEvent
2019-11-04 12:48:37
emsec/hal
https://api.github.com/repos/emsec/hal
closed
Feature: Allow manual selection of module colors
Priority: Medium Status: Abandoned Type: Feature Request
HAL sometimes selects two colors very close to each other when creating many modules, making the different modules hard to see. I suggest a manual override for that color by: - passing the color in the python command for creating a module - right-clicking on a module in the navigation view
1.0
Feature: Allow manual selection of module colors - HAL sometimes selects two colors very close to each other when creating many modules, making the different modules hard to see. I suggest a manual override for that color by: - passing the color in the python command for creating a module - right-clicking on a module in the navigation view
non_defect
feature allow manual selection of module colors hal sometimes selects two colors very close to each other when creating many modules making the different modules hard to see i suggest a manual override for that color by passing the color in the python command for creating a module right clicking on a module in the navigation view
0
51,251
13,207,402,164
IssuesEvent
2020-08-14 22:58:07
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
environment should be checked for PATH, etc. clobbering at cmake time (Trac #97)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/97">https://code.icecube.wisc.edu/projects/icecube/ticket/97</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "_ts": "1194753078000000", "description": "", "reporter": "troy", "cc": "", "resolution": "fixed", "time": "2007-08-23T23:27:36", "component": "cmake", "summary": "environment should be checked for PATH, etc. clobbering at cmake time", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
environment should be checked for PATH, etc. clobbering at cmake time (Trac #97) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/97">https://code.icecube.wisc.edu/projects/icecube/ticket/97</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "_ts": "1194753078000000", "description": "", "reporter": "troy", "cc": "", "resolution": "fixed", "time": "2007-08-23T23:27:36", "component": "cmake", "summary": "environment should be checked for PATH, etc. clobbering at cmake time", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
environment should be checked for path etc clobbering at cmake time trac migrated from json status closed changetime ts description reporter troy cc resolution fixed time component cmake summary environment should be checked for path etc clobbering at cmake time priority normal keywords milestone owner troy type defect
1
171,191
27,081,820,294
IssuesEvent
2023-02-14 14:30:58
hypha-dao/hypha_wallet
https://api.github.com/repos/hypha-dao/hypha_wallet
opened
UI review
Design
Dark mode & Light mode - [ ] Scan QR banner - [ ] Memo in transaction history (and detail) - [ ] BottomNav menu: can we have more box-shadow? (see here) - [ ] Transaction signing - action title - [ ] transaction expires in - [ ] Handle “while” sliding. - [ ] Bottom Sheets - [ ] Former Bottom sheet “transaction fail” - [ ] Import account - [ ] toggle::Active use Main app colour gradient (not green), and use the same component all over the app (also fix size pls if possible)
1.0
UI review - Dark mode & Light mode - [ ] Scan QR banner - [ ] Memo in transaction history (and detail) - [ ] BottomNav menu: can we have more box-shadow? (see here) - [ ] Transaction signing - action title - [ ] transaction expires in - [ ] Handle “while” sliding. - [ ] Bottom Sheets - [ ] Former Bottom sheet “transaction fail” - [ ] Import account - [ ] toggle::Active use Main app colour gradient (not green), and use the same component all over the app (also fix size pls if possible)
non_defect
ui review dark mode light mode scan qr banner memo in transaction history and detail bottomnav menu can we have more box shadow see here transaction signing action title transaction expires in handle “while” sliding bottom sheets former bottom sheet “transaction fail” import account toggle active use main app colour gradient not green and use the same component all over the app also fix size pls if possible
0
33,717
7,201,510,337
IssuesEvent
2018-02-05 22:54:02
otros-systems/otroslogviewer
https://api.github.com/repos/otros-systems/otroslogviewer
closed
OutOfMemory on continues tailing on big files.
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Use stock OLV 1.2, choose "tail log with type detection". Choose log4j log that is continuesly growing. 2. Wait until it reaches heap limit. 3. When it reaches memory threshold, application crashes with "java.lang.OutOfMemoryError: Java heap ...". Sometimes even error boxes are not drawed. (so cannot include exact error currently), application is unusable from this point. Nothing at screen gets repainted, nothing reacts to clicks. What is the expected output? What do you see instead? a) Expected output usual application behavior tailing logs. (preffered) b) Grid with log lines cleared to free up memory. Notification is shown somewhere (which would NOT require to click ok to proceed) about error cause. This needed when log watching is left for overnight. It should not crash. What version of the product are you using? On what operating system? OLV 1.2, java version "1.7.0_67" Java(TM) SE Runtime Environment (build 1.7.0_67-b01) Java HotSpot(TM) 64-Bit Server VM (build 24.65-b04, mixed mode) Windows 7x64 (Microsoft Windows [Version 6.1.7601]) Please provide any additional information below. Problem could be repeated on Windows Server 20xx OS. Desired fix is to be able to select option in settings: default size of "current log window". That is number of lines that are actually created in grid and occupies memory. That is if window is 1000 lines, then when it comes 1002 line, first two lines (1,2) are deleted from "current window" and freed from memory. if i use scroll bar to go to the begining of the file, then "current window" is lines 1-1000, if i go to the last lines, than "current window" becomes 3-1002 lines. ``` Original issue reported on code.google.com by `jaun...@gmail.com` on 5 Nov 2014 at 7:30
1.0
OutOfMemory on continues tailing on big files. - ``` What steps will reproduce the problem? 1. Use stock OLV 1.2, choose "tail log with type detection". Choose log4j log that is continuesly growing. 2. Wait until it reaches heap limit. 3. When it reaches memory threshold, application crashes with "java.lang.OutOfMemoryError: Java heap ...". Sometimes even error boxes are not drawed. (so cannot include exact error currently), application is unusable from this point. Nothing at screen gets repainted, nothing reacts to clicks. What is the expected output? What do you see instead? a) Expected output usual application behavior tailing logs. (preffered) b) Grid with log lines cleared to free up memory. Notification is shown somewhere (which would NOT require to click ok to proceed) about error cause. This needed when log watching is left for overnight. It should not crash. What version of the product are you using? On what operating system? OLV 1.2, java version "1.7.0_67" Java(TM) SE Runtime Environment (build 1.7.0_67-b01) Java HotSpot(TM) 64-Bit Server VM (build 24.65-b04, mixed mode) Windows 7x64 (Microsoft Windows [Version 6.1.7601]) Please provide any additional information below. Problem could be repeated on Windows Server 20xx OS. Desired fix is to be able to select option in settings: default size of "current log window". That is number of lines that are actually created in grid and occupies memory. That is if window is 1000 lines, then when it comes 1002 line, first two lines (1,2) are deleted from "current window" and freed from memory. if i use scroll bar to go to the begining of the file, then "current window" is lines 1-1000, if i go to the last lines, than "current window" becomes 3-1002 lines. ``` Original issue reported on code.google.com by `jaun...@gmail.com` on 5 Nov 2014 at 7:30
defect
outofmemory on continues tailing on big files what steps will reproduce the problem use stock olv choose tail log with type detection choose log that is continuesly growing wait until it reaches heap limit when it reaches memory threshold application crashes with java lang outofmemoryerror java heap sometimes even error boxes are not drawed so cannot include exact error currently application is unusable from this point nothing at screen gets repainted nothing reacts to clicks what is the expected output what do you see instead a expected output usual application behavior tailing logs preffered b grid with log lines cleared to free up memory notification is shown somewhere which would not require to click ok to proceed about error cause this needed when log watching is left for overnight it should not crash what version of the product are you using on what operating system olv java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode windows microsoft windows please provide any additional information below problem could be repeated on windows server os desired fix is to be able to select option in settings default size of current log window that is number of lines that are actually created in grid and occupies memory that is if window is lines then when it comes line first two lines are deleted from current window and freed from memory if i use scroll bar to go to the begining of the file then current window is lines if i go to the last lines than current window becomes lines original issue reported on code google com by jaun gmail com on nov at
1
62,400
17,023,915,141
IssuesEvent
2021-07-03 04:32:08
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
missing boundaries for continents in nominatim search
Component: nominatim Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 1.41pm, Tuesday, 16th December 2014]** when searching for australia, europe... e.g. [http://nominatim.openstreetmap.org/search.php?q=Europe&polygon=1] it finds them on the map but does not show boundaries.
1.0
missing boundaries for continents in nominatim search - **[Submitted to the original trac issue database at 1.41pm, Tuesday, 16th December 2014]** when searching for australia, europe... e.g. [http://nominatim.openstreetmap.org/search.php?q=Europe&polygon=1] it finds them on the map but does not show boundaries.
defect
missing boundaries for continents in nominatim search when searching for australia europe e g it finds them on the map but does not show boundaries
1
176,059
14,549,172,041
IssuesEvent
2020-12-16 03:07:09
liamlau/individual-project
https://api.github.com/repos/liamlau/individual-project
opened
Write Documentation Requirements Wiki Page
documentation
At https://github.com/liamlau/individual-project/wiki/Documentation-%7C-Requirements-(MoSCoW) Take the OneNote notes I have and write them here in a better format.
1.0
Write Documentation Requirements Wiki Page - At https://github.com/liamlau/individual-project/wiki/Documentation-%7C-Requirements-(MoSCoW) Take the OneNote notes I have and write them here in a better format.
non_defect
write documentation requirements wiki page at take the onenote notes i have and write them here in a better format
0
73,393
24,607,520,654
IssuesEvent
2022-10-14 17:43:19
idaholab/HERON
https://api.github.com/repos/idaholab/HERON
closed
[DEFECT] Mac Documentation Failed
defect
-------- Defect Description -------- **Describe the defect** ##### What did you expect to see happen? The user manual to be built ##### What did you see instead? Import errors stopping the build process <img width="1016" alt="image" src="https://user-images.githubusercontent.com/74023858/195851747-b024752a-fdb7-43f4-ac34-7ded3aeb8e9c.png"> ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. Update and build raven heron and teal 2. Activate raven_libraries environment 3. Try to build HERON documentation 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or task? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[DEFECT] Mac Documentation Failed - -------- Defect Description -------- **Describe the defect** ##### What did you expect to see happen? The user manual to be built ##### What did you see instead? Import errors stopping the build process <img width="1016" alt="image" src="https://user-images.githubusercontent.com/74023858/195851747-b024752a-fdb7-43f4-ac34-7ded3aeb8e9c.png"> ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. Update and build raven heron and teal 2. Activate raven_libraries environment 3. Try to build HERON documentation 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or task? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
defect
mac documentation failed defect description describe the defect what did you expect to see happen the user manual to be built what did you see instead import errors stopping the build process img width alt image src do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior update and build raven heron and teal activate raven libraries environment try to build heron documentation screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
1
2,437
3,683,179,959
IssuesEvent
2016-02-24 13:03:08
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
support for letsencrypt
enhancement infrastructure security related
http://letsencrypt.org is awesome and would allow us to (hopefully) automate certificate granting
1.0
support for letsencrypt - http://letsencrypt.org is awesome and would allow us to (hopefully) automate certificate granting
non_defect
support for letsencrypt is awesome and would allow us to hopefully automate certificate granting
0
174,937
21,300,580,151
IssuesEvent
2022-04-15 02:11:15
turkdevops/vue-devtools
https://api.github.com/repos/turkdevops/vue-devtools
opened
CVE-2021-43138 (High) detected in async-2.6.1.tgz
security vulnerability
## CVE-2021-43138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-2.6.1.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.1.tgz">https://registry.npmjs.org/async/-/async-2.6.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/async/package.json</p> <p> Dependency Hierarchy: - cypress-3.8.3.tgz (Root Library) - getos-3.1.1.tgz - :x: **async-2.6.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>dependabot/npm_and_yarn/electron-9.4.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution (async): 3.2.2</p> <p>Direct dependency fix Resolution (cypress): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-43138 (High) detected in async-2.6.1.tgz - ## CVE-2021-43138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-2.6.1.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.1.tgz">https://registry.npmjs.org/async/-/async-2.6.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/async/package.json</p> <p> Dependency Hierarchy: - cypress-3.8.3.tgz (Root Library) - getos-3.1.1.tgz - :x: **async-2.6.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>dependabot/npm_and_yarn/electron-9.4.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution (async): 3.2.2</p> <p>Direct dependency fix Resolution (cypress): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in async tgz cve high severity vulnerability vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file package json path to vulnerable library node modules async package json dependency hierarchy cypress tgz root library getos tgz x async tgz vulnerable library found in base branch dependabot npm and yarn electron vulnerability details a vulnerability exists in async through fixed in which could let a malicious user obtain privileges via the mapvalues method publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async direct dependency fix resolution cypress step up your open source security game with whitesource
0
20,580
3,385,468,587
IssuesEvent
2015-11-27 11:41:57
contao/core
https://api.github.com/repos/contao/core
closed
\Request and relative 301 location
defect
Requesting entries from a sitemap http://www.heise.de/sitemap.xml Those entries are redirected to a relative URL, which is wrong handled by the Request::send function Array ( [scheme] => http [host] => www.heise.de [path] => /meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html ) Redirect goes to : Array ( [path] => /newsticker/meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html ) that means, the redirect here (https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Request.php#L443) has to check, if scheme, host, port exists and rebuild the URL
1.0
\Request and relative 301 location - Requesting entries from a sitemap http://www.heise.de/sitemap.xml Those entries are redirected to a relative URL, which is wrong handled by the Request::send function Array ( [scheme] => http [host] => www.heise.de [path] => /meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html ) Redirect goes to : Array ( [path] => /newsticker/meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html ) that means, the redirect here (https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Request.php#L443) has to check, if scheme, host, port exists and rebuild the URL
defect
request and relative location requesting entries from a sitemap those entries are redirected to a relative url which is wrong handled by the request send function array http meldung hickhack um virtual reality apps fuer samsungs gear vr html redirect goes to array newsticker meldung hickhack um virtual reality apps fuer samsungs gear vr html that means the redirect here has to check if scheme host port exists and rebuild the url
1
61,056
17,023,590,485
IssuesEvent
2021-07-03 02:48:47
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
OSM Server Side Script: sketch_line and sketch_route produce erratic output
Component: utils Priority: major Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 11.22pm, Friday, 14th May 2010]** sketch_line (sketch_route is also affected) often produces erratic graphics. I have observed two issues: 1. apparently, if one of the stop members of the relation is an area tagged as amenity=bus_station, it is omitted completely. For example: [http://78.46.81.38/api/sketch-line?network=ATM&ref=80&style=padua] consisting of relations: [http://www.openstreetmap.org/browse/relation/611446] [http://www.openstreetmap.org/browse/relation/611538] The first stop (Bonola) is tagged as described above and missing from the sketch. This seems incorrect to me as, to my knowledge, bus_station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms. 2. Trolleybus lines (route=trolleybus) are not rendered at all - attempting to sketch them results in "No relation found". To reproduce: [http://78.46.81.38/api/sketch-line?network=ATM&ref=92&style=padua] relations: [http://www.openstreetmap.org/browse/relation/913880] [http://www.openstreetmap.org/browse/relation/913907] 3. in some cases incorrect stop names are displayed - names are mixed up or the same stop name is displayed multiple times. The graph itself seems correct (though I have not verified this in too much detail). Mild example: [http://78.46.81.38/api/sketch-line?network=ATM&ref=19&style=padua] relations: [http://www.openstreetmap.org/browse/relation/661028] [http://www.openstreetmap.org/browse/relation/661000] Here just one stop name is wrong: the 11th (from Roserio) is displayed as "Duomo" when in fact it should be "Piazzale Accursio". (Duomo is the last stop.) Stops at which only one direction stops seem to be particular prone to this error (though it happens to all). A more extreme one would be: [http://78.46.81.38/api/sketch-line?network=ATM&ref=78&style=padua] relations: [http://www.openstreetmap.org/browse/relation/420649] [http://www.openstreetmap.org/browse/relation/445904] This one is botched beyond recognition and usability: names are either missing completely, or the same name is repeated over and over multiple times. And finally: [http://78.46.81.38/api/sketch-line?network=ATM&ref=69&style=padua] two alternative routes, resulting in a total of four relations: [http://www.openstreetmap.org/browse/relation/536964] [http://www.openstreetmap.org/browse/relation/537040] [http://www.openstreetmap.org/browse/relation/536992] [http://www.openstreetmap.org/browse/relation/537008]
1.0
OSM Server Side Script: sketch_line and sketch_route produce erratic output - **[Submitted to the original trac issue database at 11.22pm, Friday, 14th May 2010]** sketch_line (sketch_route is also affected) often produces erratic graphics. I have observed two issues: 1. apparently, if one of the stop members of the relation is an area tagged as amenity=bus_station, it is omitted completely. For example: [http://78.46.81.38/api/sketch-line?network=ATM&ref=80&style=padua] consisting of relations: [http://www.openstreetmap.org/browse/relation/611446] [http://www.openstreetmap.org/browse/relation/611538] The first stop (Bonola) is tagged as described above and missing from the sketch. This seems incorrect to me as, to my knowledge, bus_station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms. 2. Trolleybus lines (route=trolleybus) are not rendered at all - attempting to sketch them results in "No relation found". To reproduce: [http://78.46.81.38/api/sketch-line?network=ATM&ref=92&style=padua] relations: [http://www.openstreetmap.org/browse/relation/913880] [http://www.openstreetmap.org/browse/relation/913907] 3. in some cases incorrect stop names are displayed - names are mixed up or the same stop name is displayed multiple times. The graph itself seems correct (though I have not verified this in too much detail). Mild example: [http://78.46.81.38/api/sketch-line?network=ATM&ref=19&style=padua] relations: [http://www.openstreetmap.org/browse/relation/661028] [http://www.openstreetmap.org/browse/relation/661000] Here just one stop name is wrong: the 11th (from Roserio) is displayed as "Duomo" when in fact it should be "Piazzale Accursio". (Duomo is the last stop.) Stops at which only one direction stops seem to be particular prone to this error (though it happens to all). A more extreme one would be: [http://78.46.81.38/api/sketch-line?network=ATM&ref=78&style=padua] relations: [http://www.openstreetmap.org/browse/relation/420649] [http://www.openstreetmap.org/browse/relation/445904] This one is botched beyond recognition and usability: names are either missing completely, or the same name is repeated over and over multiple times. And finally: [http://78.46.81.38/api/sketch-line?network=ATM&ref=69&style=padua] two alternative routes, resulting in a total of four relations: [http://www.openstreetmap.org/browse/relation/536964] [http://www.openstreetmap.org/browse/relation/537040] [http://www.openstreetmap.org/browse/relation/536992] [http://www.openstreetmap.org/browse/relation/537008]
defect
osm server side script sketch line and sketch route produce erratic output sketch line sketch route is also affected often produces erratic graphics i have observed two issues apparently if one of the stop members of the relation is an area tagged as amenity bus station it is omitted completely for example consisting of relations the first stop bonola is tagged as described above and missing from the sketch this seems incorrect to me as to my knowledge bus station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms trolleybus lines route trolleybus are not rendered at all attempting to sketch them results in no relation found to reproduce relations in some cases incorrect stop names are displayed names are mixed up or the same stop name is displayed multiple times the graph itself seems correct though i have not verified this in too much detail mild example relations here just one stop name is wrong the from roserio is displayed as duomo when in fact it should be piazzale accursio duomo is the last stop stops at which only one direction stops seem to be particular prone to this error though it happens to all a more extreme one would be relations this one is botched beyond recognition and usability names are either missing completely or the same name is repeated over and over multiple times and finally two alternative routes resulting in a total of four relations
1
66,076
19,914,757,229
IssuesEvent
2022-01-25 21:10:59
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Space panel scroll gradient only appears after scrolling
T-Defect
### Steps to reproduce 1. Be in lots of spaces 2. Open the app 3. Note lack of gradient 4. Scroll 5. Note gradient appears 6. Scroll all the way to the top (similar to app start) 7. Note gradient is still present ### Outcome #### What did you expect? The gradient should be consistently shown. This gradient: ![image](https://user-images.githubusercontent.com/1190097/151060321-c5ee9b02-db77-4168-a9e5-fa6e5f306373.png) ### Operating system Windows 10 ### Application version Nightly (2022-01-25) ### How did you install the app? The Internet ### Homeserver t2l.io ### Will you send logs? No
1.0
Space panel scroll gradient only appears after scrolling - ### Steps to reproduce 1. Be in lots of spaces 2. Open the app 3. Note lack of gradient 4. Scroll 5. Note gradient appears 6. Scroll all the way to the top (similar to app start) 7. Note gradient is still present ### Outcome #### What did you expect? The gradient should be consistently shown. This gradient: ![image](https://user-images.githubusercontent.com/1190097/151060321-c5ee9b02-db77-4168-a9e5-fa6e5f306373.png) ### Operating system Windows 10 ### Application version Nightly (2022-01-25) ### How did you install the app? The Internet ### Homeserver t2l.io ### Will you send logs? No
defect
space panel scroll gradient only appears after scrolling steps to reproduce be in lots of spaces open the app note lack of gradient scroll note gradient appears scroll all the way to the top similar to app start note gradient is still present outcome what did you expect the gradient should be consistently shown this gradient operating system windows application version nightly how did you install the app the internet homeserver io will you send logs no
1
28,764
5,352,638,032
IssuesEvent
2017-02-20 00:19:07
AtlasOfLivingAustralia/biocache-service
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-service
closed
Download documentation - needs an edit
Downloads priority-low status-new type-defect
_From @nickdos on January 20, 2015 2:50_ _From @mbohun on August 19, 2014 11:16_ _migrated from:_ https://code.google.com/p/ala/issues/detail?id=230 _date:_ Mon Aug 12 03:24:21 2013 _author:_ johntan...@gmail.com --- Records downloaded through the BioCache come with a README file which contains a link to information about the Download Fields: [https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0](https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0) Some of the fields have explanations that are not helpful. For example: Coordinate Precision The precision of the coordinates Coordinate Uncertainty In Meters Distance from the occurrence Although there is a link to the Darwin Core Terms for the fields, the description could be more helpful for a person needing to understand what they have downloaded. Other editing requirements ACRONYMS need explaining eg GUID, IBRA, IMCRA Spelling Links to further assistance Some issue flags missing definitions _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#39_ _Copied from original issue: AtlasOfLivingAustralia/bie-service#2_
1.0
Download documentation - needs an edit - _From @nickdos on January 20, 2015 2:50_ _From @mbohun on August 19, 2014 11:16_ _migrated from:_ https://code.google.com/p/ala/issues/detail?id=230 _date:_ Mon Aug 12 03:24:21 2013 _author:_ johntan...@gmail.com --- Records downloaded through the BioCache come with a README file which contains a link to information about the Download Fields: [https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0](https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0) Some of the fields have explanations that are not helpful. For example: Coordinate Precision The precision of the coordinates Coordinate Uncertainty In Meters Distance from the occurrence Although there is a link to the Darwin Core Terms for the fields, the description could be more helpful for a person needing to understand what they have downloaded. Other editing requirements ACRONYMS need explaining eg GUID, IBRA, IMCRA Spelling Links to further assistance Some issue flags missing definitions _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#39_ _Copied from original issue: AtlasOfLivingAustralia/bie-service#2_
defect
download documentation needs an edit from nickdos on january from mbohun on august migrated from date mon aug author johntan gmail com records downloaded through the biocache come with a readme file which contains a link to information about the download fields some of the fields have explanations that are not helpful for example coordinate precision the precision of the coordinates coordinate uncertainty in meters distance from the occurrence although there is a link to the darwin core terms for the fields the description could be more helpful for a person needing to understand what they have downloaded other editing requirements acronyms need explaining eg guid ibra imcra spelling links to further assistance some issue flags missing definitions copied from original issue atlasoflivingaustralia biocache hubs copied from original issue atlasoflivingaustralia bie service
1
72,433
24,116,397,240
IssuesEvent
2022-09-20 15:01:40
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
objects sometimes create an unneeded L1 block
Type: Defect
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | nixos Distribution Version | 22.05 Kernel Version | 5.10.81 Architecture | x86-64 OpenZFS Version | 2.1.1-1 ### Describe the problem you're observing some objects are using 2 L0 block pointers directly in the dnode with no bonus data other objects are using 1 L1 pointer in the dnode with 176 bytes of bonus data, and the L1 then contains 2 L0's, but the 2 L0's + bonus could have all fit within the dnode https://github.com/openzfs/zfs/blob/master/include/sys/dnode.h#L241-L254 i think the root of the problem, is that this doesnt describe how to layout a dnode with 2 pointers, when the bonus is small enough to allow it an example object with 2 L0's, no bonus, and no L1 ``` Object lvl iblk dblk dsize dnsize lsize %full type 1 1 128K 16K 18K 512 32K 100.00 object directory (K=inherit) (Z=inherit=on) dnode flags: USED_BYTES dnode maxblkid: 1 Fat ZAP stats: Indirect blocks: 0 L0 DVA[0]=<0:4800:800> DVA[1]=<0:2004800:800> DVA[2]=<0:10008800:800> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/800P birth=4L/4P fill=1 cksum=12f05825 4000 L0 DVA[0]=<0:5000:1000> DVA[1]=<0:2005000:1000> DVA[2]=<0:10009000:1000> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/1000P birth=4L/4P fill=1 cksum=9e86 ``` an example object with a fatzap, the `.pack` file was 1 byte over the size limit of a microzap it is stored as an L1 in the dnode, with 2 L0's, and 176 bytes of bouns but the dnode could have held both L0's ``` Object lvl iblk dblk dsize dnsize lsize %full type 38 2 128K 16K 16K 512 32K 100.00 ZFS directory (K=inherit) (Z=inherit=uncompressed) 176 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED dnode maxblkid: 1 path /lk/.git/objects/pack uid 1000 gid 100 ... Fat ZAP stats: ... pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.idx = 41 (type: Regular File) pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.pack = 256 (type: Regular File) Indirect blocks: 0 L1 DVA[0]=<0:42cf000:800> DVA[1]=<0:6015800:800> [L1 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=20000L/800P birth=20L/20P fill=2 cksum=d9559b8fb0ee03b5:819a23eca9f5b5 0 L0 DVA[0]=<0:818e000:800> DVA[1]=<0:a01a000:800> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/800P birth=20L/20P fill=1 cksum=5479b2c3a5a78dd8:3a4dad7bd5e005b 4000 L0 DVA[0]=<0:818e800:1000> DVA[1]=<0:a000000:1000> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/1000P birth=20L/20P fill=1 cksum=d37a5477b8ae53c4:343b2c00b22a ``` ### Describe how to reproduce the problem this test pool was just the result of running git clone on https://github.com/littlekernel/lk but it should be possible to reproduce it on any basic file with 2 records
1.0
objects sometimes create an unneeded L1 block - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | nixos Distribution Version | 22.05 Kernel Version | 5.10.81 Architecture | x86-64 OpenZFS Version | 2.1.1-1 ### Describe the problem you're observing some objects are using 2 L0 block pointers directly in the dnode with no bonus data other objects are using 1 L1 pointer in the dnode with 176 bytes of bonus data, and the L1 then contains 2 L0's, but the 2 L0's + bonus could have all fit within the dnode https://github.com/openzfs/zfs/blob/master/include/sys/dnode.h#L241-L254 i think the root of the problem, is that this doesnt describe how to layout a dnode with 2 pointers, when the bonus is small enough to allow it an example object with 2 L0's, no bonus, and no L1 ``` Object lvl iblk dblk dsize dnsize lsize %full type 1 1 128K 16K 18K 512 32K 100.00 object directory (K=inherit) (Z=inherit=on) dnode flags: USED_BYTES dnode maxblkid: 1 Fat ZAP stats: Indirect blocks: 0 L0 DVA[0]=<0:4800:800> DVA[1]=<0:2004800:800> DVA[2]=<0:10008800:800> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/800P birth=4L/4P fill=1 cksum=12f05825 4000 L0 DVA[0]=<0:5000:1000> DVA[1]=<0:2005000:1000> DVA[2]=<0:10009000:1000> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/1000P birth=4L/4P fill=1 cksum=9e86 ``` an example object with a fatzap, the `.pack` file was 1 byte over the size limit of a microzap it is stored as an L1 in the dnode, with 2 L0's, and 176 bytes of bouns but the dnode could have held both L0's ``` Object lvl iblk dblk dsize dnsize lsize %full type 38 2 128K 16K 16K 512 32K 100.00 ZFS directory (K=inherit) (Z=inherit=uncompressed) 176 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED dnode maxblkid: 1 path /lk/.git/objects/pack uid 1000 gid 100 ... Fat ZAP stats: ... pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.idx = 41 (type: Regular File) pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.pack = 256 (type: Regular File) Indirect blocks: 0 L1 DVA[0]=<0:42cf000:800> DVA[1]=<0:6015800:800> [L1 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=20000L/800P birth=20L/20P fill=2 cksum=d9559b8fb0ee03b5:819a23eca9f5b5 0 L0 DVA[0]=<0:818e000:800> DVA[1]=<0:a01a000:800> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/800P birth=20L/20P fill=1 cksum=5479b2c3a5a78dd8:3a4dad7bd5e005b 4000 L0 DVA[0]=<0:818e800:1000> DVA[1]=<0:a000000:1000> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/1000P birth=20L/20P fill=1 cksum=d37a5477b8ae53c4:343b2c00b22a ``` ### Describe how to reproduce the problem this test pool was just the result of running git clone on https://github.com/littlekernel/lk but it should be possible to reproduce it on any basic file with 2 records
defect
objects sometimes create an unneeded block system information type version name distribution name nixos distribution version kernel version architecture openzfs version describe the problem you re observing some objects are using block pointers directly in the dnode with no bonus data other objects are using pointer in the dnode with bytes of bonus data and the then contains s but the s bonus could have all fit within the dnode i think the root of the problem is that this doesnt describe how to layout a dnode with pointers when the bonus is small enough to allow it an example object with s no bonus and no object lvl iblk dblk dsize dnsize lsize full type object directory k inherit z inherit on dnode flags used bytes dnode maxblkid fat zap stats indirect blocks dva dva dva unencrypted le contiguous unique triple size birth fill cksum dva dva dva unencrypted le contiguous unique triple size birth fill cksum an example object with a fatzap the pack file was byte over the size limit of a microzap it is stored as an in the dnode with s and bytes of bouns but the dnode could have held both s object lvl iblk dblk dsize dnsize lsize full type zfs directory k inherit z inherit uncompressed bonus system attributes dnode flags used bytes userused accounted userobjused accounted dnode maxblkid path lk git objects pack uid gid fat zap stats pack idx type regular file pack pack type regular file indirect blocks dva dva unencrypted le contiguous unique double size birth fill cksum dva dva unencrypted le contiguous unique double size birth fill cksum dva dva unencrypted le contiguous unique double size birth fill cksum describe how to reproduce the problem this test pool was just the result of running git clone on but it should be possible to reproduce it on any basic file with records
1
187,545
22,045,797,556
IssuesEvent
2022-05-30 01:27:02
utopikkad/my-Todo-List
https://api.github.com/repos/utopikkad/my-Todo-List
closed
CVE-2021-37701 (High) detected in tar-4.4.13.tgz, tar-4.4.1.tgz - autoclosed
security vulnerability
## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b></p></summary> <p> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p> Dependency Hierarchy: - cli-7.3.9.tgz (Root Library) - pacote-9.4.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - chokidar-2.0.4.tgz - fsevents-1.2.4.tgz - node-pre-gyp-0.10.0.tgz - :x: **tar-4.4.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/utopikkad/my-Todo-List/commit/bd19af551d149e2bce7801819589f1a69368d63a">bd19af551d149e2bce7801819589f1a69368d63a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.16</p> <p>Direct dependency fix Resolution (@angular/cli): 7.3.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37701 (High) detected in tar-4.4.13.tgz, tar-4.4.1.tgz - autoclosed - ## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b></p></summary> <p> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p> Dependency Hierarchy: - cli-7.3.9.tgz (Root Library) - pacote-9.4.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - chokidar-2.0.4.tgz - fsevents-1.2.4.tgz - node-pre-gyp-0.10.0.tgz - :x: **tar-4.4.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/utopikkad/my-Todo-List/commit/bd19af551d149e2bce7801819589f1a69368d63a">bd19af551d149e2bce7801819589f1a69368d63a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.16</p> <p>Direct dependency fix Resolution (@angular/cli): 7.3.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in tar tgz tar tgz autoclosed cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar tgz tar for node library home page a href dependency hierarchy cli tgz root library pacote tgz x tar tgz vulnerable library tar tgz tar for node library home page a href dependency hierarchy karma tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution angular cli step up your open source security game with whitesource
0
78,172
27,355,813,306
IssuesEvent
2023-02-27 12:48:04
matrix-org/synapse
https://api.github.com/repos/matrix-org/synapse
opened
Initial syncs being cached means you can get an outdated view of account data
A-Sync S-Minor T-Defect O-Uncommon
There are two problems with this: - (1) it possibly betrays client expectations, that doing an initial sync will get them to 'correct' state immediately. - (2) it means we don't apply the correct set of ignored users to the events received in the timeline of the initial sync. (For context: the server hides events sent by ignored users, when sending events to the client.) This issue is causing https://github.com/vector-im/element-ios/issues/7387 — we *could* tell them to just do a zero-timeout incremental sync immediately but that feels like a hack and it wouldn't solve (2). For (2), I wonder if we can just fetch the latest account data stream ID for the ignored users list, when doing an initial sync, and embedding it into the sync response cache token. That way, the sync response cache entry will automatically be invalidated when the account data changes. It's not a perfect solution since it only helps with ignored users, but it at least solves (2). Problem (1) might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data. Sliding sync might obviate these concerns anyway.
1.0
Initial syncs being cached means you can get an outdated view of account data - There are two problems with this: - (1) it possibly betrays client expectations, that doing an initial sync will get them to 'correct' state immediately. - (2) it means we don't apply the correct set of ignored users to the events received in the timeline of the initial sync. (For context: the server hides events sent by ignored users, when sending events to the client.) This issue is causing https://github.com/vector-im/element-ios/issues/7387 — we *could* tell them to just do a zero-timeout incremental sync immediately but that feels like a hack and it wouldn't solve (2). For (2), I wonder if we can just fetch the latest account data stream ID for the ignored users list, when doing an initial sync, and embedding it into the sync response cache token. That way, the sync response cache entry will automatically be invalidated when the account data changes. It's not a perfect solution since it only helps with ignored users, but it at least solves (2). Problem (1) might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data. Sliding sync might obviate these concerns anyway.
defect
initial syncs being cached means you can get an outdated view of account data there are two problems with this it possibly betrays client expectations that doing an initial sync will get them to correct state immediately it means we don t apply the correct set of ignored users to the events received in the timeline of the initial sync for context the server hides events sent by ignored users when sending events to the client this issue is causing — we could tell them to just do a zero timeout incremental sync immediately but that feels like a hack and it wouldn t solve for i wonder if we can just fetch the latest account data stream id for the ignored users list when doing an initial sync and embedding it into the sync response cache token that way the sync response cache entry will automatically be invalidated when the account data changes it s not a perfect solution since it only helps with ignored users but it at least solves problem might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data sliding sync might obviate these concerns anyway
1
814,247
30,496,367,288
IssuesEvent
2023-07-18 11:07:14
ufosc/Club_Website_2
https://api.github.com/repos/ufosc/Club_Website_2
closed
V0.2 BETA
help wanted discussion high priority
## V0.2 Beta Release Issue Tracker | Issue | Ref | Status | Assignee | |-------|-----|--------|----------| | TEST: /api/edit ROUTE | #158 | **HELP WANTED** | Thinura | | ~TEST: /api/blog route~ | #157 | DONE | @zeim839 | | ~TEST: /api/image route~ | N/A | DONE | @zeim839 | Goals: Finish testing before releasing beta to public.
1.0
V0.2 BETA - ## V0.2 Beta Release Issue Tracker | Issue | Ref | Status | Assignee | |-------|-----|--------|----------| | TEST: /api/edit ROUTE | #158 | **HELP WANTED** | Thinura | | ~TEST: /api/blog route~ | #157 | DONE | @zeim839 | | ~TEST: /api/image route~ | N/A | DONE | @zeim839 | Goals: Finish testing before releasing beta to public.
non_defect
beta beta release issue tracker issue ref status assignee test api edit route help wanted thinura test api blog route done test api image route n a done goals finish testing before releasing beta to public
0
31,799
6,626,437,349
IssuesEvent
2017-09-22 19:33:15
wheat32/Armold
https://api.github.com/repos/wheat32/Armold
closed
Close sensors at end of runtime
defect medium priority
Add a method to close all the sensors at the end of runtime. This can be done by calling `Sensor.close();`.
1.0
Close sensors at end of runtime - Add a method to close all the sensors at the end of runtime. This can be done by calling `Sensor.close();`.
defect
close sensors at end of runtime add a method to close all the sensors at the end of runtime this can be done by calling sensor close
1
158,594
12,419,653,383
IssuesEvent
2020-05-23 07:34:10
the-canonizer/canonizer.2.0
https://api.github.com/repos/the-canonizer/canonizer.2.0
closed
My Supports : Camp names are missing and getting page crash on remove support of the blank camp name
18th release Fixed Show Stopper bug ready to test
Steps to reproduce : 1. Go to https://staging.canonizer.com/ 2. Login with valid email and password 3. Go to browse 4. Select Namespace as "Sandbox testing" and select as of filter "as of date" = "6/18/2019, 5:30:00 AM" 5. Go to Account settings -> My Supports Please refer attached screenshot ![image](https://user-images.githubusercontent.com/40416610/81470689-b3f8a800-9209-11ea-8653-1784b60d1ed8.png) I think filter values are getting saved in session.User supports page also not listing all the supported camps Please confirm..
1.0
My Supports : Camp names are missing and getting page crash on remove support of the blank camp name - Steps to reproduce : 1. Go to https://staging.canonizer.com/ 2. Login with valid email and password 3. Go to browse 4. Select Namespace as "Sandbox testing" and select as of filter "as of date" = "6/18/2019, 5:30:00 AM" 5. Go to Account settings -> My Supports Please refer attached screenshot ![image](https://user-images.githubusercontent.com/40416610/81470689-b3f8a800-9209-11ea-8653-1784b60d1ed8.png) I think filter values are getting saved in session.User supports page also not listing all the supported camps Please confirm..
non_defect
my supports camp names are missing and getting page crash on remove support of the blank camp name steps to reproduce go to login with valid email and password go to browse select namespace as sandbox testing and select as of filter as of date am go to account settings my supports please refer attached screenshot i think filter values are getting saved in session user supports page also not listing all the supported camps please confirm
0
146,551
19,406,174,645
IssuesEvent
2021-12-20 01:14:18
keanhankins/ranger
https://api.github.com/repos/keanhankins/ranger
opened
CVE-2021-4104 (High) detected in log4j-1.2.17.jar
security vulnerability
## CVE-2021-4104 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: ranger/ranger-hive-plugin-shim/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2021-12-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/ranger-hive-plugin-shim/pom.xml","/ranger-examples/plugin-sampleapp/pom.xml","/plugin-nifi/pom.xml","/ranger-kms-plugin-shim/pom.xml","/ranger-kylin-plugin-shim/pom.xml","/intg/pom.xml","/ranger-hbase-plugin-shim/pom.xml","/agents-cred/pom.xml","/ranger-knox-plugin-shim/pom.xml","/hdfs-agent/pom.xml","/ranger-examples/sampleapp/pom.xml","/agents-audit/pom.xml","/agents-common/pom.xml","/security-admin/pom.xml","/ranger-presto-plugin-shim/pom.xml","/ranger-atlas-plugin-shim/pom.xml","/ranger-examples/conditions-enrichers/pom.xml","/plugin-presto/pom.xml","/ranger-ozone-plugin-shim/pom.xml","/plugin-ozone/pom.xml","/ranger-kafka-plugin-shim/pom.xml","/distro/pom.xml","/ugsync/pom.xml","/ranger-tools/pom.xml","/plugin-nifi-registry/pom.xml","/kms/pom.xml","/plugin-atlas/pom.xml","/plugin-kms/pom.xml","/plugin-kylin/pom.xml","/plugin-yarn/pom.xml","/plugin-kudu/pom.xml","/plugin-solr/pom.xml","/tagsync/pom.xml","/ranger-elasticsearch-plugin-shim/pom.xml","/plugin-elasticsearch/pom.xml","/storm-agent/pom.xml","/ranger-solr-plugin-shim/pom.xml","/ranger-hdfs-plugin-shim/pom.xml","/unixauthclient/pom.xml","/hbase-agent/pom.xml","/ranger-storm-plugin-shim/pom.xml","/ranger-yarn-plugin-shim/pom.xml","/hive-agent/pom.xml","/plugin-kafka/pom.xml","/credentialbuilder/pom.xml","/embeddedwebserver/pom.xml","/ranger-examples/sample-client/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-4104 (High) detected in log4j-1.2.17.jar - ## CVE-2021-4104 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: ranger/ranger-hive-plugin-shim/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2021-12-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/ranger-hive-plugin-shim/pom.xml","/ranger-examples/plugin-sampleapp/pom.xml","/plugin-nifi/pom.xml","/ranger-kms-plugin-shim/pom.xml","/ranger-kylin-plugin-shim/pom.xml","/intg/pom.xml","/ranger-hbase-plugin-shim/pom.xml","/agents-cred/pom.xml","/ranger-knox-plugin-shim/pom.xml","/hdfs-agent/pom.xml","/ranger-examples/sampleapp/pom.xml","/agents-audit/pom.xml","/agents-common/pom.xml","/security-admin/pom.xml","/ranger-presto-plugin-shim/pom.xml","/ranger-atlas-plugin-shim/pom.xml","/ranger-examples/conditions-enrichers/pom.xml","/plugin-presto/pom.xml","/ranger-ozone-plugin-shim/pom.xml","/plugin-ozone/pom.xml","/ranger-kafka-plugin-shim/pom.xml","/distro/pom.xml","/ugsync/pom.xml","/ranger-tools/pom.xml","/plugin-nifi-registry/pom.xml","/kms/pom.xml","/plugin-atlas/pom.xml","/plugin-kms/pom.xml","/plugin-kylin/pom.xml","/plugin-yarn/pom.xml","/plugin-kudu/pom.xml","/plugin-solr/pom.xml","/tagsync/pom.xml","/ranger-elasticsearch-plugin-shim/pom.xml","/plugin-elasticsearch/pom.xml","/storm-agent/pom.xml","/ranger-solr-plugin-shim/pom.xml","/ranger-hdfs-plugin-shim/pom.xml","/unixauthclient/pom.xml","/hbase-agent/pom.xml","/ranger-storm-plugin-shim/pom.xml","/ranger-yarn-plugin-shim/pom.xml","/hive-agent/pom.xml","/plugin-kafka/pom.xml","/credentialbuilder/pom.xml","/embeddedwebserver/pom.xml","/ranger-examples/sample-client/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file ranger ranger hive plugin shim pom xml path to vulnerable library home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar canner repository jar canner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in base branch master vulnerability details jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions vulnerabilityurl
0
202,171
23,061,486,630
IssuesEvent
2022-07-25 10:19:59
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 106: envoy-1.17.3: 4 advisories [8.3]
1.severity: security
[search](https://search.nix.gsc.io/?q=envoy&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=envoy+in%3Apath&type=Code) * [ ] [CVE-2021-32777](https://nvd.nist.gov/vuln/detail/CVE-2021-32777) CVSSv3=8.3 (nixos-unstable) * [ ] [CVE-2021-32779](https://nvd.nist.gov/vuln/detail/CVE-2021-32779) CVSSv3=8.3 (nixos-unstable) * [ ] [CVE-2021-32778](https://nvd.nist.gov/vuln/detail/CVE-2021-32778) CVSSv3=7.5 (nixos-unstable) * [ ] [CVE-2021-32781](https://nvd.nist.gov/vuln/detail/CVE-2021-32781) CVSSv3=7.5 (nixos-unstable) ## CVE details ### CVE-2021-32777 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions when ext-authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the HTTP spec. However, only the last header value is sent. This may allow specifically crafted requests to bypass authorization. Attackers may be able to escalate privileges when using ext-authz extension or back end service that uses multiple value headers for authorization. A specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext-authz extension. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to the ext-authz extension to correctly merge multiple request header values, when sending request for authorization. ### CVE-2021-32779 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy incorrectly handled a URI '#fragment' element as part of the path element. Envoy is configured with an RBAC filter for authorization or similar mechanism with an explicit case of a final "/admin" path element, or is using a negative assertion with final path element of "/admin". The client sends request to "/app1/admin#foo". In Envoy prior to 1.18.0, or 1.18.0+ configured with path_normalization=false. Envoy treats fragment as a suffix of the query string when present, or as a suffix of the path when query string is absent, so it evaluates the final path element as "/admin#foo" and mismatches with the configured "/admin" path element. In Envoy 1.18.0+ configured with path_normalization=true. Envoy transforms this to /app1/admin%23foo and mismatches with the configured /admin prefix. The resulting URI is sent to the next server-agent with the offending "#foo" fragment which violates RFC3986 or with the nonsensical "%23foo" text appended. A specifically constructed request with URI containing '#fragment' element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of Privileges when path based request authorization extensions. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes that removes fragment from URI path in incoming requests. ### CVE-2021-32778 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy’s procedure for resetting a HTTP/2 stream has O(N^2) complexity, leading to high CPU utilization when a large number of streams are reset. Deployments are susceptible to Denial of Service when Envoy is configured with high limit on H/2 concurrent streams. An attacker wishing to exploit this vulnerability would require a client opening and closing a large number of H/2 streams. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to reduce time complexity of resetting HTTP/2 streams. As a workaround users may limit the number of simultaneous HTTP/2 dreams for upstream and downstream peers to a low number, i.e. 100. ### CVE-2021-32781 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions after Envoy sends a locally generated response it must stop further processing of request or response data. However when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block. A specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a Denial of Service when using extensions that modify and increase the size of request or response bodies, such as decompressor filter. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to address incomplete termination of request processing after locally generated response. As a workaround disable Envoy's decompressor, json-transcoder or grpc-web extensions or proprietary extensions that modify and increase the size of request or response bodies, if feasible. ----- Scanned versions: nixos-unstable: bc9b956714e. Cc @lukegb
True
Vulnerability roundup 106: envoy-1.17.3: 4 advisories [8.3] - [search](https://search.nix.gsc.io/?q=envoy&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=envoy+in%3Apath&type=Code) * [ ] [CVE-2021-32777](https://nvd.nist.gov/vuln/detail/CVE-2021-32777) CVSSv3=8.3 (nixos-unstable) * [ ] [CVE-2021-32779](https://nvd.nist.gov/vuln/detail/CVE-2021-32779) CVSSv3=8.3 (nixos-unstable) * [ ] [CVE-2021-32778](https://nvd.nist.gov/vuln/detail/CVE-2021-32778) CVSSv3=7.5 (nixos-unstable) * [ ] [CVE-2021-32781](https://nvd.nist.gov/vuln/detail/CVE-2021-32781) CVSSv3=7.5 (nixos-unstable) ## CVE details ### CVE-2021-32777 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions when ext-authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the HTTP spec. However, only the last header value is sent. This may allow specifically crafted requests to bypass authorization. Attackers may be able to escalate privileges when using ext-authz extension or back end service that uses multiple value headers for authorization. A specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext-authz extension. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to the ext-authz extension to correctly merge multiple request header values, when sending request for authorization. ### CVE-2021-32779 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy incorrectly handled a URI '#fragment' element as part of the path element. Envoy is configured with an RBAC filter for authorization or similar mechanism with an explicit case of a final "/admin" path element, or is using a negative assertion with final path element of "/admin". The client sends request to "/app1/admin#foo". In Envoy prior to 1.18.0, or 1.18.0+ configured with path_normalization=false. Envoy treats fragment as a suffix of the query string when present, or as a suffix of the path when query string is absent, so it evaluates the final path element as "/admin#foo" and mismatches with the configured "/admin" path element. In Envoy 1.18.0+ configured with path_normalization=true. Envoy transforms this to /app1/admin%23foo and mismatches with the configured /admin prefix. The resulting URI is sent to the next server-agent with the offending "#foo" fragment which violates RFC3986 or with the nonsensical "%23foo" text appended. A specifically constructed request with URI containing '#fragment' element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of Privileges when path based request authorization extensions. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes that removes fragment from URI path in incoming requests. ### CVE-2021-32778 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy’s procedure for resetting a HTTP/2 stream has O(N^2) complexity, leading to high CPU utilization when a large number of streams are reset. Deployments are susceptible to Denial of Service when Envoy is configured with high limit on H/2 concurrent streams. An attacker wishing to exploit this vulnerability would require a client opening and closing a large number of H/2 streams. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to reduce time complexity of resetting HTTP/2 streams. As a workaround users may limit the number of simultaneous HTTP/2 dreams for upstream and downstream peers to a low number, i.e. 100. ### CVE-2021-32781 Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions after Envoy sends a locally generated response it must stop further processing of request or response data. However when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block. A specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a Denial of Service when using extensions that modify and increase the size of request or response bodies, such as decompressor filter. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to address incomplete termination of request processing after locally generated response. As a workaround disable Envoy's decompressor, json-transcoder or grpc-web extensions or proprietary extensions that modify and increase the size of request or response bodies, if feasible. ----- Scanned versions: nixos-unstable: bc9b956714e. Cc @lukegb
non_defect
vulnerability roundup envoy advisories nixos unstable nixos unstable nixos unstable nixos unstable cve details cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions when ext authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the http spec however only the last header value is sent this may allow specifically crafted requests to bypass authorization attackers may be able to escalate privileges when using ext authz extension or back end service that uses multiple value headers for authorization a specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext authz extension envoy versions contain fixes to the ext authz extension to correctly merge multiple request header values when sending request for authorization cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions envoy incorrectly handled a uri fragment element as part of the path element envoy is configured with an rbac filter for authorization or similar mechanism with an explicit case of a final admin path element or is using a negative assertion with final path element of admin the client sends request to admin foo in envoy prior to or configured with path normalization false envoy treats fragment as a suffix of the query string when present or as a suffix of the path when query string is absent so it evaluates the final path element as admin foo and mismatches with the configured admin path element in envoy configured with path normalization true envoy transforms this to admin and mismatches with the configured admin prefix the resulting uri is sent to the next server agent with the offending foo fragment which violates or with the nonsensical text appended a specifically constructed request with uri containing fragment element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of privileges when path based request authorization extensions envoy versions contain fixes that removes fragment from uri path in incoming requests cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions envoy’s procedure for resetting a http stream has o n complexity leading to high cpu utilization when a large number of streams are reset deployments are susceptible to denial of service when envoy is configured with high limit on h concurrent streams an attacker wishing to exploit this vulnerability would require a client opening and closing a large number of h streams envoy versions contain fixes to reduce time complexity of resetting http streams as a workaround users may limit the number of simultaneous http dreams for upstream and downstream peers to a low number i e cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions after envoy sends a locally generated response it must stop further processing of request or response data however when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block a specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a denial of service when using extensions that modify and increase the size of request or response bodies such as decompressor filter envoy versions contain fixes to address incomplete termination of request processing after locally generated response as a workaround disable envoy s decompressor json transcoder or grpc web extensions or proprietary extensions that modify and increase the size of request or response bodies if feasible scanned versions nixos unstable cc lukegb
0
27,930
5,412,428,396
IssuesEvent
2017-03-01 14:32:27
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
Unespected behaviour drawing arc with ellipse()
bug category: documentation
#### System information (version) - OpenCV => 3.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2015 #### Detailed description Both doc and code works different than expected. It looks that `startAngle` and `endAngle` refer to the projections on the ellipse of the wanted arc on the bounding circle. Look at the image ![ellipse](https://cloud.githubusercontent.com/assets/15944968/23079369/fc7c73b6-f54c-11e6-977b-bf344f223876.png) The blue arc has been generated 30/45/135 as angle/start/end. The red arc has been using same angles as above but choosing equal axis. May be this is wanted but is not same as is in [the doc](http://docs.opencv.org/3.2.0/d6/d6e/group__imgproc__draw.html#ga28b2267d35786f5f890ca167236cbc69). #### Steps to reproduce ```.cpp int TestEllipse() { int angle = 30,startAngle = 45,endAngle = 135; Mat img(400, 400, CV_8UC3, Scalar(255, 255, 255)); Point center = img.size() / 2; int major = 0.4*img.rows; Size sz(major, major*0.5); line(img, Point(center.x, 0), Point(center.x, img.rows), Scalar(0), 1); line(img, Point(0, center.y), Point(img.cols, center.y), Scalar(0), 1); ellipse(img, center, sz, angle, 0, 360, CL_GREEN, 1); ellipse(img, center, sz, angle, startAngle, endAngle, CL_BLUE, 2); sz = Size(sz.width, sz.width); ellipse(img, center, sz, angle, 0, 360, CL_MAGENTA, 1); ellipse(img, center, sz, angle, startAngle, endAngle, CL_RED, 2); vector<double> mags, angles, x, y; double mag = 1.2 * major, alpha = angle; mags = { mag, mag }; angles = { (alpha + startAngle), (alpha + endAngle) }; polarToCart(mags, angles, x, y, true); for (size_t i = 0; i < x.size(); i++) line(img, center, center + Point(x[i], y[i]), Scalar(160, 160, 160)); imshow("ellipse", img); waitKey(0); return 0; } ```
1.0
Unespected behaviour drawing arc with ellipse() - #### System information (version) - OpenCV => 3.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2015 #### Detailed description Both doc and code works different than expected. It looks that `startAngle` and `endAngle` refer to the projections on the ellipse of the wanted arc on the bounding circle. Look at the image ![ellipse](https://cloud.githubusercontent.com/assets/15944968/23079369/fc7c73b6-f54c-11e6-977b-bf344f223876.png) The blue arc has been generated 30/45/135 as angle/start/end. The red arc has been using same angles as above but choosing equal axis. May be this is wanted but is not same as is in [the doc](http://docs.opencv.org/3.2.0/d6/d6e/group__imgproc__draw.html#ga28b2267d35786f5f890ca167236cbc69). #### Steps to reproduce ```.cpp int TestEllipse() { int angle = 30,startAngle = 45,endAngle = 135; Mat img(400, 400, CV_8UC3, Scalar(255, 255, 255)); Point center = img.size() / 2; int major = 0.4*img.rows; Size sz(major, major*0.5); line(img, Point(center.x, 0), Point(center.x, img.rows), Scalar(0), 1); line(img, Point(0, center.y), Point(img.cols, center.y), Scalar(0), 1); ellipse(img, center, sz, angle, 0, 360, CL_GREEN, 1); ellipse(img, center, sz, angle, startAngle, endAngle, CL_BLUE, 2); sz = Size(sz.width, sz.width); ellipse(img, center, sz, angle, 0, 360, CL_MAGENTA, 1); ellipse(img, center, sz, angle, startAngle, endAngle, CL_RED, 2); vector<double> mags, angles, x, y; double mag = 1.2 * major, alpha = angle; mags = { mag, mag }; angles = { (alpha + startAngle), (alpha + endAngle) }; polarToCart(mags, angles, x, y, true); for (size_t i = 0; i < x.size(); i++) line(img, center, center + Point(x[i], y[i]), Scalar(160, 160, 160)); imshow("ellipse", img); waitKey(0); return 0; } ```
non_defect
unespected behaviour drawing arc with ellipse system information version opencv operating system platform windows bit compiler visual studio detailed description both doc and code works different than expected it looks that startangle and endangle refer to the projections on the ellipse of the wanted arc on the bounding circle look at the image the blue arc has been generated as angle start end the red arc has been using same angles as above but choosing equal axis may be this is wanted but is not same as is in steps to reproduce cpp int testellipse int angle startangle endangle mat img cv scalar point center img size int major img rows size sz major major line img point center x point center x img rows scalar line img point center y point img cols center y scalar ellipse img center sz angle cl green ellipse img center sz angle startangle endangle cl blue sz size sz width sz width ellipse img center sz angle cl magenta ellipse img center sz angle startangle endangle cl red vector mags angles x y double mag major alpha angle mags mag mag angles alpha startangle alpha endangle polartocart mags angles x y true for size t i i x size i line img center center point x y scalar imshow ellipse img waitkey return
0
450,415
31,899,771,720
IssuesEvent
2023-09-18 06:53:21
jacksonhooi/INF2001_P5-4
https://api.github.com/repos/jacksonhooi/INF2001_P5-4
closed
#9 Identify Non-Functional User Requirements
documentation
Task Description: Discuss with the group and identify non-functional user requirements. Initial Requirements (from project brief) 1. The app should be Web-based in a language of your choosing 2. The manager should be able to visualise the staff workload immediately on the landing page 3. The manager should be able to allocate jobs to staff for one week at a time 4. The manager should be able to view up to three staff availability and any relevant information to make the job assignment easier on the job allocation page 5. When displaying the staff availability, the workload assigned, staff’s job preference, staff’s location at a particular date, and availabilities for the week should be shown 6. On the manager’s landing page, the top three staff with the lowest workload should be shown, and highlight all staff over 40 hours of jobs allocated 7. Staff should be able to view their weekly job assignments and overall workload for the month on their landing page 8. Staff can add and edit their availabilities up to 5 weeks ahead of time. 9. Staff can indicate their job preference for the week 10. Staff can reject jobs assigned to them, but they will be warned to discuss the jobs with their manager before proceeding with the rejection 11. The company’s IT administrators will oversee adding new staff and managers to the system Goals: - [x] List of non-functional requirements created Completion Criteria Successful creation of list of non functional requirements Allocated Time: 10 Days Start date: 5 Sept End date: 15 Sept
1.0
#9 Identify Non-Functional User Requirements - Task Description: Discuss with the group and identify non-functional user requirements. Initial Requirements (from project brief) 1. The app should be Web-based in a language of your choosing 2. The manager should be able to visualise the staff workload immediately on the landing page 3. The manager should be able to allocate jobs to staff for one week at a time 4. The manager should be able to view up to three staff availability and any relevant information to make the job assignment easier on the job allocation page 5. When displaying the staff availability, the workload assigned, staff’s job preference, staff’s location at a particular date, and availabilities for the week should be shown 6. On the manager’s landing page, the top three staff with the lowest workload should be shown, and highlight all staff over 40 hours of jobs allocated 7. Staff should be able to view their weekly job assignments and overall workload for the month on their landing page 8. Staff can add and edit their availabilities up to 5 weeks ahead of time. 9. Staff can indicate their job preference for the week 10. Staff can reject jobs assigned to them, but they will be warned to discuss the jobs with their manager before proceeding with the rejection 11. The company’s IT administrators will oversee adding new staff and managers to the system Goals: - [x] List of non-functional requirements created Completion Criteria Successful creation of list of non functional requirements Allocated Time: 10 Days Start date: 5 Sept End date: 15 Sept
non_defect
identify non functional user requirements task description discuss with the group and identify non functional user requirements initial requirements from project brief the app should be web based in a language of your choosing the manager should be able to visualise the staff workload immediately on the landing page the manager should be able to allocate jobs to staff for one week at a time the manager should be able to view up to three staff availability and any relevant information to make the job assignment easier on the job allocation page when displaying the staff availability the workload assigned staff’s job preference staff’s location at a particular date and availabilities for the week should be shown on the manager’s landing page the top three staff with the lowest workload should be shown and highlight all staff over hours of jobs allocated staff should be able to view their weekly job assignments and overall workload for the month on their landing page staff can add and edit their availabilities up to weeks ahead of time staff can indicate their job preference for the week staff can reject jobs assigned to them but they will be warned to discuss the jobs with their manager before proceeding with the rejection the company’s it administrators will oversee adding new staff and managers to the system goals list of non functional requirements created completion criteria successful creation of list of non functional requirements allocated time days start date sept end date sept
0
37,124
8,215,468,696
IssuesEvent
2018-09-05 05:41:04
masteroy/algorithm
https://api.github.com/repos/masteroy/algorithm
closed
[LeetCode] 150. Evaluate Reverse Polish Notation
LeetCode Medium
Evaluate the value of an arithmetic expression in Reverse Polish Notation. Valid operators are +, -, *, /. Each operand may be an integer or another expression. **Note:** - The division between two integers should truncate toward zero. - The given RPN expression is always valid. That means the expression would always evaluate to a result and there won't be any divide by zero operation. **Example 1:** ``` Input: ["2", "1", "+", "3", "*"] Output: 9 Explanation: ((2 + 1) * 3) = 9 ``` **Example 2:** ``` Input: ["4", "13", "5", "/", "+"] Output: 6 Explanation: (4 + (13 / 5)) = 6 ``` **Example 3:** ``` Input: ["10", "6", "9", "3", "+", "-11", "*", "/", "*", "17", "+", "5", "+"] Output: 22 Explanation: ((10 * (6 / ((9 + 3) * -11))) + 17) + 5 = ((10 * (6 / (12 * -11))) + 17) + 5 = ((10 * (6 / -132)) + 17) + 5 = ((10 * 0) + 17) + 5 = (0 + 17) + 5 = 17 + 5 = 22 ```
1.0
[LeetCode] 150. Evaluate Reverse Polish Notation - Evaluate the value of an arithmetic expression in Reverse Polish Notation. Valid operators are +, -, *, /. Each operand may be an integer or another expression. **Note:** - The division between two integers should truncate toward zero. - The given RPN expression is always valid. That means the expression would always evaluate to a result and there won't be any divide by zero operation. **Example 1:** ``` Input: ["2", "1", "+", "3", "*"] Output: 9 Explanation: ((2 + 1) * 3) = 9 ``` **Example 2:** ``` Input: ["4", "13", "5", "/", "+"] Output: 6 Explanation: (4 + (13 / 5)) = 6 ``` **Example 3:** ``` Input: ["10", "6", "9", "3", "+", "-11", "*", "/", "*", "17", "+", "5", "+"] Output: 22 Explanation: ((10 * (6 / ((9 + 3) * -11))) + 17) + 5 = ((10 * (6 / (12 * -11))) + 17) + 5 = ((10 * (6 / -132)) + 17) + 5 = ((10 * 0) + 17) + 5 = (0 + 17) + 5 = 17 + 5 = 22 ```
non_defect
evaluate reverse polish notation evaluate the value of an arithmetic expression in reverse polish notation valid operators are each operand may be an integer or another expression note the division between two integers should truncate toward zero the given rpn expression is always valid that means the expression would always evaluate to a result and there won t be any divide by zero operation example input output explanation example input output explanation example input output explanation
0
1,625
2,603,968,448
IssuesEvent
2015-02-24 18:59:39
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳阴茎里面有疙瘩
auto-migrated Priority-Medium Type-Defect
``` 沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17
1.0
沈阳阴茎里面有疙瘩 - ``` 沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17
defect
沈阳阴茎里面有疙瘩 沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
1
134,612
12,623,614,780
IssuesEvent
2020-06-14 00:10:01
CoffeePerry/mercury-py
https://api.github.com/repos/CoffeePerry/mercury-py
opened
Load first documentation
documentation
# Load first documentation Load to [README.md](https://github.com/CoffeePerry/mercury-py/blob/master/README.md): - Prerequisites; - First configurations; - Future; - More info. Load all to [wiki](https://github.com/CoffeePerry/mercury-py/wiki).
1.0
Load first documentation - # Load first documentation Load to [README.md](https://github.com/CoffeePerry/mercury-py/blob/master/README.md): - Prerequisites; - First configurations; - Future; - More info. Load all to [wiki](https://github.com/CoffeePerry/mercury-py/wiki).
non_defect
load first documentation load first documentation load to prerequisites first configurations future more info load all to
0
257,565
19,523,916,316
IssuesEvent
2021-12-30 01:53:41
BlazorComponent/MASA.Blazor
https://api.github.com/repos/BlazorComponent/MASA.Blazor
closed
【UI组件】Text fields - 示例,清空/修改文本框的内容后,点击空白区域,文本框内容被重置
bug documentation resolved
隐藏详细信息 - Another input ,输入文本后,点击空白区域,输入的内容被清空 ![文本框 - 隐藏详细信息 ,文本框输入文本后点击空白区域文本被清空](https://user-images.githubusercontent.com/95004531/146323631-b3088d90-9a47-4900-a513-9f10d6aa97f5.png)
1.0
【UI组件】Text fields - 示例,清空/修改文本框的内容后,点击空白区域,文本框内容被重置 - 隐藏详细信息 - Another input ,输入文本后,点击空白区域,输入的内容被清空 ![文本框 - 隐藏详细信息 ,文本框输入文本后点击空白区域文本被清空](https://user-images.githubusercontent.com/95004531/146323631-b3088d90-9a47-4900-a513-9f10d6aa97f5.png)
non_defect
【ui组件】text fields 示例,清空 修改文本框的内容后,点击空白区域,文本框内容被重置 隐藏详细信息 another input 输入文本后,点击空白区域,输入的内容被清空
0
776,686
27,264,560,939
IssuesEvent
2023-02-22 17:03:11
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
opened
CP-2(8): Contingency Plan | Identify Critical Assets
Priority: P3 Suggested Assignment: IT Security Function ITSG-33 Class: Operational Control: CP-2
# Control Definition CONTINGENCY PLAN | IDENTIFY CRITICAL ASSETS The organization identifies critical information system assets supporting essential missions and business functions. # Class Operational # Supplemental Guidance Organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including, for example, as part of business impact analyses. Organizations identify critical information system assets so that additional safeguards and countermeasures can be employed (above and beyond those safeguards and countermeasures routinely implemented) to help ensure that organizational missions/business functions can continue to be conducted during contingency operations. In addition, the identification of critical information assets facilitates the prioritization of organizational resources. Critical information system assets include technical and operational aspects. Technical aspects include, for example, information technology services, information system components, information technology products, and mechanisms. Operational aspects include, for example, procedures (manually executed operations) and personnel (individuals operating technical safeguards and/or executing manual procedures). Organizational program protection plans can provide assistance in identifying critical assets. Related controls: SA-14, SA-15. # Suggested Assignment IT Security Function # Support Teams IT Operations Group
1.0
CP-2(8): Contingency Plan | Identify Critical Assets - # Control Definition CONTINGENCY PLAN | IDENTIFY CRITICAL ASSETS The organization identifies critical information system assets supporting essential missions and business functions. # Class Operational # Supplemental Guidance Organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including, for example, as part of business impact analyses. Organizations identify critical information system assets so that additional safeguards and countermeasures can be employed (above and beyond those safeguards and countermeasures routinely implemented) to help ensure that organizational missions/business functions can continue to be conducted during contingency operations. In addition, the identification of critical information assets facilitates the prioritization of organizational resources. Critical information system assets include technical and operational aspects. Technical aspects include, for example, information technology services, information system components, information technology products, and mechanisms. Operational aspects include, for example, procedures (manually executed operations) and personnel (individuals operating technical safeguards and/or executing manual procedures). Organizational program protection plans can provide assistance in identifying critical assets. Related controls: SA-14, SA-15. # Suggested Assignment IT Security Function # Support Teams IT Operations Group
non_defect
cp contingency plan identify critical assets control definition contingency plan identify critical assets the organization identifies critical information system assets supporting essential missions and business functions class operational supplemental guidance organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including for example as part of business impact analyses organizations identify critical information system assets so that additional safeguards and countermeasures can be employed above and beyond those safeguards and countermeasures routinely implemented to help ensure that organizational missions business functions can continue to be conducted during contingency operations in addition the identification of critical information assets facilitates the prioritization of organizational resources critical information system assets include technical and operational aspects technical aspects include for example information technology services information system components information technology products and mechanisms operational aspects include for example procedures manually executed operations and personnel individuals operating technical safeguards and or executing manual procedures organizational program protection plans can provide assistance in identifying critical assets related controls sa sa suggested assignment it security function support teams it operations group
0
307,159
26,518,545,168
IssuesEvent
2023-01-18 23:19:01
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_graph_break_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass)
module: flaky-tests skipped module: unknown
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_graph_break_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472). Over the past 72 hours, it has flakily failed in 2 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_graph_break_dynamic_shapes` Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
1.0
DISABLED test_graph_break_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_graph_break_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472). Over the past 72 hours, it has flakily failed in 2 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_graph_break_dynamic_shapes` Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
non_defect
disabled test graph break dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test graph break dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode
0
73,767
7,354,580,655
IssuesEvent
2018-03-09 07:38:10
FullScreenShenanigans/FullScreenPokemon
https://api.github.com/repos/FullScreenShenanigans/FullScreenPokemon
opened
Unit test gaining experience in battle
Good First Issue Help Wanted Test Coverage
When an enemy Pokemon is defeated, it should give experience to the Pokemon that defeated it. Some suggestions of cases to test (or file separate issues for if they're not yet implemented): * One of your Pokemon defeated it * >=2 of your Pokemon defeated it * Exp share * Exp all
1.0
Unit test gaining experience in battle - When an enemy Pokemon is defeated, it should give experience to the Pokemon that defeated it. Some suggestions of cases to test (or file separate issues for if they're not yet implemented): * One of your Pokemon defeated it * >=2 of your Pokemon defeated it * Exp share * Exp all
non_defect
unit test gaining experience in battle when an enemy pokemon is defeated it should give experience to the pokemon that defeated it some suggestions of cases to test or file separate issues for if they re not yet implemented one of your pokemon defeated it of your pokemon defeated it exp share exp all
0
33,662
2,770,762,849
IssuesEvent
2015-05-01 16:53:05
Metaswitch/gemini
https://api.github.com/repos/Metaswitch/gemini
closed
g.3gpp.ics media feature tag rendering is wrong
bug cat:easy medium-priority
When used in Accept-Contact/Contact/Reject-Contact headers, media tags are rendered with a leading ‘+’: Accept-Contact: *;+g.3gpp.ics Gemini does not include the ‘+’, it should.
1.0
g.3gpp.ics media feature tag rendering is wrong - When used in Accept-Contact/Contact/Reject-Contact headers, media tags are rendered with a leading ‘+’: Accept-Contact: *;+g.3gpp.ics Gemini does not include the ‘+’, it should.
non_defect
g ics media feature tag rendering is wrong when used in accept contact contact reject contact headers media tags are rendered with a leading ‘ ’ accept contact g ics gemini does not include the ‘ ’ it should
0
22,177
3,609,714,283
IssuesEvent
2016-02-05 00:11:45
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
dart2js_extra/deferred_custom_loader_test fails in csp mode on browsers
area-dart2js Type-Defect
The test dart2js_extra/deferred_custom_loader_test, located at tests/compiler/dart2js_extra/deferred_custom_loader_test.dart, is failing when run on linux content_shell in CSP mode, proabably for obvious reasons. If this is expected, then the status should be changed to SkipByDesign, as it is for the deferred_fail_and_retry_test. I'm doing that now, marking it SkipByDesign, with this issue number cited. So if this is as expected, and not going to change, just close the issue and leave the status unchanged. Here is the error: FAILED: dart2js-drt-checked release_ia32 dart2js_extra/deferred_custom_loader_test Expected: Pass Actual: RuntimeError CommandOutput[dart2js]: CONSOLE ERROR: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: "script-src 'self'". CONSOLE MESSAGE: line 78: { "type": "window_onerror", "value": "window.onerror called: \n\nhttp://127.0.0.1:52467/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_deferred_custom_loader_test/test.js:5045:\nUncaught EvalError: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: \"script-src 'self'\".\n\n\n", "timestamp": 0.087 Short reproduction command (experimental): python tools/test.py -mrelease -cdart2js -rdrt -aia32 --use-sdk --write-debug-log --write-test-outcome-log --clear_browser_cache --minified --dart2js-batch --csp --checked -t480 dart2js_extra/deferred_custom_loader_test
1.0
dart2js_extra/deferred_custom_loader_test fails in csp mode on browsers - The test dart2js_extra/deferred_custom_loader_test, located at tests/compiler/dart2js_extra/deferred_custom_loader_test.dart, is failing when run on linux content_shell in CSP mode, proabably for obvious reasons. If this is expected, then the status should be changed to SkipByDesign, as it is for the deferred_fail_and_retry_test. I'm doing that now, marking it SkipByDesign, with this issue number cited. So if this is as expected, and not going to change, just close the issue and leave the status unchanged. Here is the error: FAILED: dart2js-drt-checked release_ia32 dart2js_extra/deferred_custom_loader_test Expected: Pass Actual: RuntimeError CommandOutput[dart2js]: CONSOLE ERROR: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: "script-src 'self'". CONSOLE MESSAGE: line 78: { "type": "window_onerror", "value": "window.onerror called: \n\nhttp://127.0.0.1:52467/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_deferred_custom_loader_test/test.js:5045:\nUncaught EvalError: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: \"script-src 'self'\".\n\n\n", "timestamp": 0.087 Short reproduction command (experimental): python tools/test.py -mrelease -cdart2js -rdrt -aia32 --use-sdk --write-debug-log --write-test-outcome-log --clear_browser_cache --minified --dart2js-batch --csp --checked -t480 dart2js_extra/deferred_custom_loader_test
defect
extra deferred custom loader test fails in csp mode on browsers the test extra deferred custom loader test located at tests compiler extra deferred custom loader test dart is failing when run on linux content shell in csp mode proabably for obvious reasons if this is expected then the status should be changed to skipbydesign as it is for the deferred fail and retry test i m doing that now marking it skipbydesign with this issue number cited so if this is as expected and not going to change just close the issue and leave the status unchanged here is the error failed drt checked release extra deferred custom loader test expected pass actual runtimeerror commandoutput console error refused to evaluate a string as javascript because unsafe eval is not an allowed source of script in the following content security policy directive script src self console message line type window onerror value window onerror called n n evalerror refused to evaluate a string as javascript because unsafe eval is not an allowed source of script in the following content security policy directive script src self n n n timestamp short reproduction command experimental python tools test py mrelease rdrt use sdk write debug log write test outcome log clear browser cache minified batch csp checked extra deferred custom loader test
1
21,174
3,466,971,940
IssuesEvent
2015-12-22 08:41:03
Ryzhehvost/keyla
https://api.github.com/repos/Ryzhehvost/keyla
closed
keyla-0.1.9-x64 would not start under win7-64
auto-migrated duplicate Priority-Medium Type-Defect
``` 1. Install keyla-0.1.9-x64-setup.exe 2. Run keyla.exe Application immediately crushes with 0xc000007b keyla-0.1.9-x64, Win7 Pro 64 bit ``` Original issue reported on code.google.com by `smishc...@gmail.com` on 9 Jan 2013 at 2:24 Attachments: * [1.PNG](https://storage.googleapis.com/google-code-attachments/keyla/issue-28/comment-0/1.PNG)
1.0
keyla-0.1.9-x64 would not start under win7-64 - ``` 1. Install keyla-0.1.9-x64-setup.exe 2. Run keyla.exe Application immediately crushes with 0xc000007b keyla-0.1.9-x64, Win7 Pro 64 bit ``` Original issue reported on code.google.com by `smishc...@gmail.com` on 9 Jan 2013 at 2:24 Attachments: * [1.PNG](https://storage.googleapis.com/google-code-attachments/keyla/issue-28/comment-0/1.PNG)
defect
keyla would not start under install keyla setup exe run keyla exe application immediately crushes with keyla pro bit original issue reported on code google com by smishc gmail com on jan at attachments
1
22,864
3,727,389,257
IssuesEvent
2016-03-06 08:04:54
godfather1103/mentohust
https://api.github.com/repos/godfather1103/mentohust
closed
mipsel版mentohust认证锐捷4.85,提示“用户不允许使用本服务”
auto-migrated Priority-Medium Type-Defect
``` 使用Windows版Mentohust认证一切正常,但是使用Mipsel版用相同的� ��数和方式进行认证,会收到“用户不允许使用本服务”的提 示。 返回结果如下 ** Netcard: vlan2 ** Auth timeout(s): 8 s ** Heartbeat interval(s): 30 s ** Failure waittime(s): 15 s ** Allow failures: 8 times ** Multicast addr: Ruijie ** DHCP Type: Disabled ** MAC: 94:44:52:af:e7:64 ** IP: 0.0.0.0 ** Netmask: 0.0.0.0 !! Save the authentication parameters to /jffs/etc/mentohust.conf failure锛? !! Save the authentication parameters to /tmp/root/mentohust.conf failure锛? >> Looking for server... ** Authentication MAC: 00:1a:a9:1c:9b:c6 >> Send user name... >> Send password... >> failure! $$ System hint: 用户不允许使用本服务! 最后一次正常使用是在两个月前,锐捷的版本为4.85,并未更� �� ``` Original issue reported on code.google.com by `Xyzh...@gmail.com` on 31 Aug 2013 at 4:30
1.0
mipsel版mentohust认证锐捷4.85,提示“用户不允许使用本服务” - ``` 使用Windows版Mentohust认证一切正常,但是使用Mipsel版用相同的� ��数和方式进行认证,会收到“用户不允许使用本服务”的提 示。 返回结果如下 ** Netcard: vlan2 ** Auth timeout(s): 8 s ** Heartbeat interval(s): 30 s ** Failure waittime(s): 15 s ** Allow failures: 8 times ** Multicast addr: Ruijie ** DHCP Type: Disabled ** MAC: 94:44:52:af:e7:64 ** IP: 0.0.0.0 ** Netmask: 0.0.0.0 !! Save the authentication parameters to /jffs/etc/mentohust.conf failure锛? !! Save the authentication parameters to /tmp/root/mentohust.conf failure锛? >> Looking for server... ** Authentication MAC: 00:1a:a9:1c:9b:c6 >> Send user name... >> Send password... >> failure! $$ System hint: 用户不允许使用本服务! 最后一次正常使用是在两个月前,锐捷的版本为4.85,并未更� �� ``` Original issue reported on code.google.com by `Xyzh...@gmail.com` on 31 Aug 2013 at 4:30
defect
,提示“用户不允许使用本服务” 使用windows版mentohust认证一切正常,但是使用mipsel版用相同的� ��数和方式进行认证,会收到“用户不允许使用本服务”的提 示。 返回结果如下 netcard auth timeout s s heartbeat interval s s failure waittime s s allow failures times multicast addr ruijie dhcp type disabled mac af ip netmask save the authentication parameters to jffs etc mentohust conf failure锛 save the authentication parameters to tmp root mentohust conf failure锛 looking for server authentication mac send user name send password failure system hint 用户不允许使用本服务 最后一次正常使用是在两个月前, ,并未更� �� original issue reported on code google com by xyzh gmail com on aug at
1
59,059
17,015,345,758
IssuesEvent
2021-07-02 11:11:28
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
When viewing "My Tracks" up and down arrows affect the window behind
Component: potlatch2 Priority: minor Type: defect
**[Submitted to the original trac issue database at 2.01pm, Saturday, 22nd January 2011]** To reproduce: Open Potlatch2 Select "My Tracks" from the top left Attempt to scroll down (or up) the list with the arrow keys Nothing obvious seems to happen, but in fact the (smudged-out) map behind is moving. When the "My Tracks" box is closed, you're somewhere that you don't expect to be. If the arrow keys are supposed to move the map behind then it presumably shouldn't be smudged-out? Also some kind of clue about what the arrow keys do here would be useful. If they're not, how about making them just, er, scroll up and down the list in the foreground?
1.0
When viewing "My Tracks" up and down arrows affect the window behind - **[Submitted to the original trac issue database at 2.01pm, Saturday, 22nd January 2011]** To reproduce: Open Potlatch2 Select "My Tracks" from the top left Attempt to scroll down (or up) the list with the arrow keys Nothing obvious seems to happen, but in fact the (smudged-out) map behind is moving. When the "My Tracks" box is closed, you're somewhere that you don't expect to be. If the arrow keys are supposed to move the map behind then it presumably shouldn't be smudged-out? Also some kind of clue about what the arrow keys do here would be useful. If they're not, how about making them just, er, scroll up and down the list in the foreground?
defect
when viewing my tracks up and down arrows affect the window behind to reproduce open select my tracks from the top left attempt to scroll down or up the list with the arrow keys nothing obvious seems to happen but in fact the smudged out map behind is moving when the my tracks box is closed you re somewhere that you don t expect to be if the arrow keys are supposed to move the map behind then it presumably shouldn t be smudged out also some kind of clue about what the arrow keys do here would be useful if they re not how about making them just er scroll up and down the list in the foreground
1
59,107
17,015,744,334
IssuesEvent
2021-07-02 11:46:58
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
proposed route=bicycle relation overrides normal cycle route
Component: opencyclemap Priority: major Type: defect
**[Submitted to the original trac issue database at 9.28pm, Saturday, 5th November 2011]** http://www.openstreetmap.org/?lat=28.513&lon=-81.4106&zoom=13&layers=C http://www.openstreetmap.org/browse/way/11182207 The street south of the lake in the center is tagged lcn=yes and is part of an lcn=proposed relation. It should be displayed as normal lcn, not proposed.
1.0
proposed route=bicycle relation overrides normal cycle route - **[Submitted to the original trac issue database at 9.28pm, Saturday, 5th November 2011]** http://www.openstreetmap.org/?lat=28.513&lon=-81.4106&zoom=13&layers=C http://www.openstreetmap.org/browse/way/11182207 The street south of the lake in the center is tagged lcn=yes and is part of an lcn=proposed relation. It should be displayed as normal lcn, not proposed.
defect
proposed route bicycle relation overrides normal cycle route the street south of the lake in the center is tagged lcn yes and is part of an lcn proposed relation it should be displayed as normal lcn not proposed
1
29,279
8,316,443,215
IssuesEvent
2018-09-25 09:03:52
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
Search breaks when keybindings editor is open
broken-build important search
- Open keybindings editor - Search in search viewlet - Spins forever ``` TypeError: Cannot read property 'toString' of null errors.ts:51 at EditorService.doGetOpened (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:374:38) at EditorService.isOpen (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:335:27) at file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:298:46 at Array.forEach (<anonymous>) at SearchService.getLocalResults (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:293:24) at SearchService.search (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:97:37) at SearchModel.search (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/common/searchModel.js:743:53) at SearchView.onQueryTriggered (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:1095:28) at file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:886:23 ``` Looks like `resource` can be null here https://github.com/Microsoft/vscode/blob/master/src/vs/workbench/services/editor/browser/editorService.ts#L422 This breaks smoketests too https://dev.azure.com/vscode/VSCode/_build/results?buildId=7711&view=logs
1.0
Search breaks when keybindings editor is open - - Open keybindings editor - Search in search viewlet - Spins forever ``` TypeError: Cannot read property 'toString' of null errors.ts:51 at EditorService.doGetOpened (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:374:38) at EditorService.isOpen (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:335:27) at file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:298:46 at Array.forEach (<anonymous>) at SearchService.getLocalResults (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:293:24) at SearchService.search (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:97:37) at SearchModel.search (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/common/searchModel.js:743:53) at SearchView.onQueryTriggered (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:1095:28) at file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:886:23 ``` Looks like `resource` can be null here https://github.com/Microsoft/vscode/blob/master/src/vs/workbench/services/editor/browser/editorService.ts#L422 This breaks smoketests too https://dev.azure.com/vscode/VSCode/_build/results?buildId=7711&view=logs
non_defect
search breaks when keybindings editor is open open keybindings editor search in search viewlet spins forever typeerror cannot read property tostring of null errors ts at editorservice dogetopened file users roblou code vscode out vs workbench services editor browser editorservice js at editorservice isopen file users roblou code vscode out vs workbench services editor browser editorservice js at file users roblou code vscode out vs workbench services search node searchservice js at array foreach at searchservice getlocalresults file users roblou code vscode out vs workbench services search node searchservice js at searchservice search file users roblou code vscode out vs workbench services search node searchservice js at searchmodel search file users roblou code vscode out vs workbench parts search common searchmodel js at searchview onquerytriggered file users roblou code vscode out vs workbench parts search browser searchview js at file users roblou code vscode out vs workbench parts search browser searchview js looks like resource can be null here this breaks smoketests too
0
164,762
12,812,914,181
IssuesEvent
2020-07-04 09:36:53
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
closed
RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80
CWE-134 bug components software flawfinder flawfinder_level_4 mitigated robot component: PX4 static analysis testing triage version: v1.8.0 version: v1.9.0
```yaml id: 2720 title: 'RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80' type: bug description: If format strings can be influenced by an attacker, they can be exploited (CWE-134). Use a constant for the format specification. . Happening @ ...4_layer/px4_qurt_impl.cpp:80 cwe: - CWE-134 cve: None keywords: - flawfinder - flawfinder_level_4 - static analysis - testing - triage - CWE-134 - bug - 'version: v1.9.0' - 'robot component: PX4' - components software system: ./Firmware/platforms/qurt/src/px4_layer/px4_qurt_impl.cpp:80:2 vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2720 flaw: phase: testing specificity: subject-specific architectural-location: application-specific application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (15:05) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (15:05) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2720 reproducibility: always trace: (context) \tprintf(fmt, args); reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: Use a constant for the format specification pull-request: '' date-mitigation: '' ```
1.0
RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80 - ```yaml id: 2720 title: 'RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80' type: bug description: If format strings can be influenced by an attacker, they can be exploited (CWE-134). Use a constant for the format specification. . Happening @ ...4_layer/px4_qurt_impl.cpp:80 cwe: - CWE-134 cve: None keywords: - flawfinder - flawfinder_level_4 - static analysis - testing - triage - CWE-134 - bug - 'version: v1.9.0' - 'robot component: PX4' - components software system: ./Firmware/platforms/qurt/src/px4_layer/px4_qurt_impl.cpp:80:2 vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2720 flaw: phase: testing specificity: subject-specific architectural-location: application-specific application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (15:05) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (15:05) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2720 reproducibility: always trace: (context) \tprintf(fmt, args); reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: Use a constant for the format specification pull-request: '' date-mitigation: '' ```
non_defect
rvd cwe format if format strings can be influenced by an attacker they can be exploi layer qurt impl cpp yaml id title rvd cwe format if format strings can be influenced by an attacker they can be exploi layer qurt impl cpp type bug description if format strings can be influenced by an attacker they can be exploited cwe use a constant for the format specification happening layer qurt impl cpp cwe cwe cve none keywords flawfinder flawfinder level static analysis testing triage cwe bug version robot component components software system firmware platforms qurt src layer qurt impl cpp vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace context tprintf fmt args reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline jobs artifacts download exploitation description exploitation image exploitation vector exploitation recipe mitigation description use a constant for the format specification pull request date mitigation
0
65,756
19,680,702,156
IssuesEvent
2022-01-11 16:29:08
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
taking a video call does not always default to loudspeaker
T-Defect
### Steps to reproduce call somebody on their iphone. they're not so technologically inclined as to open this report themselves. i'll see if i can get version numbers, but it should be decently updated as iOS does. ### Outcome #### What did you expect? default to loudspeaker #### What happened instead? earphone speaker is selected by default ### Your phone model couple years old ### Operating system version _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
taking a video call does not always default to loudspeaker - ### Steps to reproduce call somebody on their iphone. they're not so technologically inclined as to open this report themselves. i'll see if i can get version numbers, but it should be decently updated as iOS does. ### Outcome #### What did you expect? default to loudspeaker #### What happened instead? earphone speaker is selected by default ### Your phone model couple years old ### Operating system version _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
taking a video call does not always default to loudspeaker steps to reproduce call somebody on their iphone they re not so technologically inclined as to open this report themselves i ll see if i can get version numbers but it should be decently updated as ios does outcome what did you expect default to loudspeaker what happened instead earphone speaker is selected by default your phone model couple years old operating system version no response application version no response homeserver no response will you send logs no
1
1,708
2,603,969,851
IssuesEvent
2015-02-24 18:59:58
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳龟头有豆豆怎么回事
auto-migrated Priority-Medium Type-Defect
``` 沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:25
1.0
沈阳龟头有豆豆怎么回事 - ``` 沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:25
defect
沈阳龟头有豆豆怎么回事 沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at
1
78,269
27,403,861,125
IssuesEvent
2023-03-01 04:17:55
pageseeder/diffx
https://api.github.com/repos/pageseeder/diffx
closed
Please add to Maven Central
Priority-Medium Type-Defect auto-migrated
``` Please add wo-diffx to Maven Central. You can follow this guide: https://docs.sonatype.org/display/Repository/Sonatype+OSS+Maven+Repository+Usage +Guide thanks .. Jason ``` Original issue reported on code.google.com by `jhar...@gmail.com` on 27 May 2013 at 1:35
1.0
Please add to Maven Central - ``` Please add wo-diffx to Maven Central. You can follow this guide: https://docs.sonatype.org/display/Repository/Sonatype+OSS+Maven+Repository+Usage +Guide thanks .. Jason ``` Original issue reported on code.google.com by `jhar...@gmail.com` on 27 May 2013 at 1:35
defect
please add to maven central please add wo diffx to maven central you can follow this guide guide thanks jason original issue reported on code google com by jhar gmail com on may at
1
21,862
3,573,522,110
IssuesEvent
2016-01-27 07:02:03
hunnee/jsontoken
https://api.github.com/repos/hunnee/jsontoken
closed
UrlBasedVerifierProvider does not work with recent version of gson
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. In pom.xml switch to a recent version of gson, 2.2.4 for example 2. Use UrlBasedVerifierProvider to fetch a certicate from an url. 3. What is the expected output? What do you see instead? I should get a List<Verifier> I get an exception [INFO] com.google.gson.JsonSyntaxException: com.google.gson.stream.MalformedJsonException: Use JsonReader.setLenient(true) to accept malformed JSON at line 2 column 2 [INFO] at com.google.gson.JsonParser.parse(JsonParser.java:65) [INFO] at com.google.gson.JsonParser.parse(JsonParser.java:45) [INFO] at net.oauth.jsontoken.discovery.UrlBasedVerifierProvider.findVerifier(UrlBasedVeri fierProvider.java:58) [INFO] at net.oauth.jsontoken.JsonTokenParser.verifyAndDeserialize(JsonTokenParser.java:10 8) How to Fix it @ https://code.google.com/p/jsontoken/source/browse/trunk/src/main/java/net/oauth/ jsontoken/discovery/UrlBasedVerifierProvider.java#51 replace String line = ""; do { line = buff.readLine(); content.append(line + "\n"); } while (line != null); with String line = buff.readLine(); while (line != null) { content.append(line + "\n"); line = buff.readLine(); } ; The later won't add an unwanted "null" string at the end of the json string. ``` Original issue reported on code.google.com by `aurelien...@gmail.com` on 22 Aug 2013 at 8:10
1.0
UrlBasedVerifierProvider does not work with recent version of gson - ``` What steps will reproduce the problem? 1. In pom.xml switch to a recent version of gson, 2.2.4 for example 2. Use UrlBasedVerifierProvider to fetch a certicate from an url. 3. What is the expected output? What do you see instead? I should get a List<Verifier> I get an exception [INFO] com.google.gson.JsonSyntaxException: com.google.gson.stream.MalformedJsonException: Use JsonReader.setLenient(true) to accept malformed JSON at line 2 column 2 [INFO] at com.google.gson.JsonParser.parse(JsonParser.java:65) [INFO] at com.google.gson.JsonParser.parse(JsonParser.java:45) [INFO] at net.oauth.jsontoken.discovery.UrlBasedVerifierProvider.findVerifier(UrlBasedVeri fierProvider.java:58) [INFO] at net.oauth.jsontoken.JsonTokenParser.verifyAndDeserialize(JsonTokenParser.java:10 8) How to Fix it @ https://code.google.com/p/jsontoken/source/browse/trunk/src/main/java/net/oauth/ jsontoken/discovery/UrlBasedVerifierProvider.java#51 replace String line = ""; do { line = buff.readLine(); content.append(line + "\n"); } while (line != null); with String line = buff.readLine(); while (line != null) { content.append(line + "\n"); line = buff.readLine(); } ; The later won't add an unwanted "null" string at the end of the json string. ``` Original issue reported on code.google.com by `aurelien...@gmail.com` on 22 Aug 2013 at 8:10
defect
urlbasedverifierprovider does not work with recent version of gson what steps will reproduce the problem in pom xml switch to a recent version of gson for example use urlbasedverifierprovider to fetch a certicate from an url what is the expected output what do you see instead i should get a list i get an exception com google gson jsonsyntaxexception com google gson stream malformedjsonexception use jsonreader setlenient true to accept malformed json at line column at com google gson jsonparser parse jsonparser java at com google gson jsonparser parse jsonparser java at net oauth jsontoken discovery urlbasedverifierprovider findverifier urlbasedveri fierprovider java at net oauth jsontoken jsontokenparser verifyanddeserialize jsontokenparser java how to fix it jsontoken discovery urlbasedverifierprovider java replace string line do line buff readline content append line n while line null with string line buff readline while line null content append line n line buff readline the later won t add an unwanted null string at the end of the json string original issue reported on code google com by aurelien gmail com on aug at
1
2,434
11,950,542,903
IssuesEvent
2020-04-03 15:22:18
gcdevops/HRWhiteListing
https://api.github.com/repos/gcdevops/HRWhiteListing
closed
Automate creation of custom fields and subfields
automation
Creation of fields needs to be automated Related to #76
1.0
Automate creation of custom fields and subfields - Creation of fields needs to be automated Related to #76
non_defect
automate creation of custom fields and subfields creation of fields needs to be automated related to
0
49,672
13,187,249,365
IssuesEvent
2020-08-13 02:49:12
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
PolyplopiaSegment issue at Detector Level (Trac #1851)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1851">https://code.icecube.wisc.edu/ticket/1851</a>, reported by saxani and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:54", "description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer", "reporter": "saxani", "cc": "", "resolution": "fixed", "_ts": "1550067174476394", "component": "combo simulation", "summary": "PolyplopiaSegment issue at Detector Level", "priority": "normal", "keywords": "", "time": "2016-09-06T14:31:56", "milestone": "", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
1.0
PolyplopiaSegment issue at Detector Level (Trac #1851) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1851">https://code.icecube.wisc.edu/ticket/1851</a>, reported by saxani and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:54", "description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer", "reporter": "saxani", "cc": "", "resolution": "fixed", "_ts": "1550067174476394", "component": "combo simulation", "summary": "PolyplopiaSegment issue at Detector Level", "priority": "normal", "keywords": "", "time": "2016-09-06T14:31:56", "milestone": "", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
defect
polyplopiasegment issue at detector level trac migrated from json status closed changetime description hi n ni m trying to run polyplopiasegment on the simulation release at detector level i end up with this error n n data user saxani environments builddet env shell sh line bus error path path ld library path ld library path dyld library path dyld library path pythonpath pythonpath icetray classpath icetray classpath platform src src build build ports ports testdata testdata rootsys cvmfs icecube opensciencegrid org rhel root shell shell new shell argv n ncan we reimplement the older version of polyplopiasegment into this release so that it can run after photon propagation n nattn juan carlos n nthanks n nspencer reporter saxani cc resolution fixed ts component combo simulation summary polyplopiasegment issue at detector level priority normal keywords time milestone owner juancarlos type defect
1
25,056
4,186,289,305
IssuesEvent
2016-06-23 14:06:07
jbs1/testingtrac
https://api.github.com/repos/jbs1/testingtrac
closed
tabbing and friends (Trac #1369)
bindings defect Incomplete Migration Migrated from Trac
Migrated from https://trac.kwarc.info/ticket/1369 ```json { "changetime": "2011-07-11T13:45:23", "cc": "miller", "blockedby": "", "keywords": "", "owner": "stamer", "due_close": "", "severity": "normal", "include_gantt": "0", "priority": "high", "due_assign": "", "type": "defect", "status": "closed", "description": "in math/papers/9901119, we find \\begin{tabbing} and\\kill, they do not work yet. \n\nI suspect they are TeX builtins", "reporter": "kohlhase", "_ts": "1310391923000000", "component": "bindings", "dependencies": "", "milestone": "", "blocking": "", "summary": "tabbing and friends", "time": "2007-01-22T17:16:45", "resolution": "fixed" } ```
1.0
tabbing and friends (Trac #1369) - Migrated from https://trac.kwarc.info/ticket/1369 ```json { "changetime": "2011-07-11T13:45:23", "cc": "miller", "blockedby": "", "keywords": "", "owner": "stamer", "due_close": "", "severity": "normal", "include_gantt": "0", "priority": "high", "due_assign": "", "type": "defect", "status": "closed", "description": "in math/papers/9901119, we find \\begin{tabbing} and\\kill, they do not work yet. \n\nI suspect they are TeX builtins", "reporter": "kohlhase", "_ts": "1310391923000000", "component": "bindings", "dependencies": "", "milestone": "", "blocking": "", "summary": "tabbing and friends", "time": "2007-01-22T17:16:45", "resolution": "fixed" } ```
defect
tabbing and friends trac migrated from json changetime cc miller blockedby keywords owner stamer due close severity normal include gantt priority high due assign type defect status closed description in math papers we find begin tabbing and kill they do not work yet n ni suspect they are tex builtins reporter kohlhase ts component bindings dependencies milestone blocking summary tabbing and friends time resolution fixed
1
1,130
2,596,584,097
IssuesEvent
2015-02-20 21:44:06
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
Compilation error with papi
category: components type: defect
``` /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: error: ‘function_nonser’ is not a member of ‘hpx::performance_counters::papi::util’ bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: note: suggested alternative: In file included from /home/pagrubel/hpx/hpx/util/function.hpp:13:0, from /home/pagrubel/hpx/hpx/hpx_fwd.hpp:43, from /home/pagrubel/hpx/hpx/hpx.hpp:9, from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15: /home/pagrubel/hpx/hpx/util/detail/function_template.hpp:343:54: note: ‘hpx::util::function_nonser’ using function_nonser = function<Sig, void, void>; ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:55: error: ‘startup_func’ was not declared in this scope bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ In file included from /usr/include/c++/4.9/bits/atomic_base.h:36:0, from /usr/include/c++/4.9/atomic:41, from /opt/boost/1.56.0-release/boost/lockfree/detail/atomic.hpp:32, from /opt/boost/1.56.0-release/boost/lockfree/queue.hpp:19, from /home/pagrubel/hpx/hpx/util/lockfree/freelist.hpp:15, from /home/pagrubel/hpx/hpx/runtime/threads/thread_data.hpp:26, from /home/pagrubel/hpx/hpx/include/threadmanager.hpp:15, from /home/pagrubel/hpx/hpx/include/runtime.hpp:14, from /home/pagrubel/hpx/hpx/hpx.hpp:11, from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15: /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:69: error: expected primary-expression before ‘bool’ bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:86: error: expression list treated as compound expression in initializer [-fpermissive] bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:289:5: error: expected ‘,’ or ‘;’ before ‘{’ token { ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input hpx::performance_counters::papi::util::get_options_description); ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input /opt/boost/1.56.0-release/boost/format/feed_args.hpp: In instantiation of ‘void boost::io::detail::call_put_head(std::basic_ostream<_CharT, _Traits>&, const void*) [with Ch = char; Tr = std::char_traits<char>; T = long int]’: /opt/boost/1.56.0-release/boost/format/feed_args.hpp:135:47: required from ‘boost::io::detail::put_holder<Ch, Tr>::put_holder(T&) [with T = long int; Ch = char; Tr = std::char_traits<char>]’ ```
1.0
Compilation error with papi - ``` /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: error: ‘function_nonser’ is not a member of ‘hpx::performance_counters::papi::util’ bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: note: suggested alternative: In file included from /home/pagrubel/hpx/hpx/util/function.hpp:13:0, from /home/pagrubel/hpx/hpx/hpx_fwd.hpp:43, from /home/pagrubel/hpx/hpx/hpx.hpp:9, from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15: /home/pagrubel/hpx/hpx/util/detail/function_template.hpp:343:54: note: ‘hpx::util::function_nonser’ using function_nonser = function<Sig, void, void>; ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:55: error: ‘startup_func’ was not declared in this scope bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ In file included from /usr/include/c++/4.9/bits/atomic_base.h:36:0, from /usr/include/c++/4.9/atomic:41, from /opt/boost/1.56.0-release/boost/lockfree/detail/atomic.hpp:32, from /opt/boost/1.56.0-release/boost/lockfree/queue.hpp:19, from /home/pagrubel/hpx/hpx/util/lockfree/freelist.hpp:15, from /home/pagrubel/hpx/hpx/runtime/threads/thread_data.hpp:26, from /home/pagrubel/hpx/hpx/include/threadmanager.hpp:15, from /home/pagrubel/hpx/hpx/include/runtime.hpp:14, from /home/pagrubel/hpx/hpx/hpx.hpp:11, from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15: /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:69: error: expected primary-expression before ‘bool’ bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:86: error: expression list treated as compound expression in initializer [-fpermissive] bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup) ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:289:5: error: expected ‘,’ or ‘;’ before ‘{’ token { ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input hpx::performance_counters::papi::util::get_options_description); ^ /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input /opt/boost/1.56.0-release/boost/format/feed_args.hpp: In instantiation of ‘void boost::io::detail::call_put_head(std::basic_ostream<_CharT, _Traits>&, const void*) [with Ch = char; Tr = std::char_traits<char>; T = long int]’: /opt/boost/1.56.0-release/boost/format/feed_args.hpp:135:47: required from ‘boost::io::detail::put_holder<Ch, Tr>::put_holder(T&) [with T = long int; Ch = char; Tr = std::char_traits<char>]’ ```
defect
compilation error with papi home pagrubel hpx src components papi counters papi startup cpp error ‘function nonser’ is not a member of ‘hpx performance counters papi util’ bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp note suggested alternative in file included from home pagrubel hpx hpx util function hpp from home pagrubel hpx hpx hpx fwd hpp from home pagrubel hpx hpx hpx hpp from home pagrubel hpx src components papi counters papi startup cpp home pagrubel hpx hpx util detail function template hpp note ‘hpx util function nonser’ using function nonser function home pagrubel hpx src components papi counters papi startup cpp error ‘startup func’ was not declared in this scope bool check startup util function nonser startup func bool pre startup in file included from usr include c bits atomic base h from usr include c atomic from opt boost release boost lockfree detail atomic hpp from opt boost release boost lockfree queue hpp from home pagrubel hpx hpx util lockfree freelist hpp from home pagrubel hpx hpx runtime threads thread data hpp from home pagrubel hpx hpx include threadmanager hpp from home pagrubel hpx hpx include runtime hpp from home pagrubel hpx hpx hpx hpp from home pagrubel hpx src components papi counters papi startup cpp home pagrubel hpx src components papi counters papi startup cpp error expected primary expression before ‘bool’ bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp error expression list treated as compound expression in initializer bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ or ‘ ’ before ‘ ’ token home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input hpx performance counters papi util get options description home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input opt boost release boost format feed args hpp in instantiation of ‘void boost io detail call put head std basic ostream const void ’ opt boost release boost format feed args hpp required from ‘boost io detail put holder put holder t ’
1
4,104
10,575,755,479
IssuesEvent
2019-10-07 16:21:22
fac-17/Autsera
https://api.github.com/repos/fac-17/Autsera
closed
Router has one endpoint for multiple instances of a component, figure out how to pass props to router endpoint.
architecture
- [x] PlacePage component is rendered on the /page endpoint, but it needs to know which place to render. - [x] it will then read values for that place from the global data array
1.0
Router has one endpoint for multiple instances of a component, figure out how to pass props to router endpoint. - - [x] PlacePage component is rendered on the /page endpoint, but it needs to know which place to render. - [x] it will then read values for that place from the global data array
non_defect
router has one endpoint for multiple instances of a component figure out how to pass props to router endpoint placepage component is rendered on the page endpoint but it needs to know which place to render it will then read values for that place from the global data array
0
77,069
26,752,554,746
IssuesEvent
2023-01-30 20:53:21
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
issue with keys
T-Defect
### Steps to reproduce 1. Where are you starting? What can you see? Bisq General room 2. What do you click? Not clicking anything. Scanning messages in the room, I can see an apparent struggle to decrypt messages going on. 3. More steps… My MacBook Pro started struggling to decrypt in an encrypted room and despite signing in and out of Element, I continue to receive the following error message "Some messages could not be decrypted Unfortunately, there are no other verified devices to request decryption keys from. Signing in and verifying other devices may help avoid this situation in the future." I left the room and tried to reenter and was urged to open an issue here due to an error. ### Outcome #### What did you expect? #### What happened instead? error message keeps reappearing. Struggle to decrypt iMessages. Final error message recited above. Messages in room prior to January 24 no longer appear. ### Operating system Mac OS 10.15.7 ### Application version Version 1.11.21-rc.1 (1.11.21-rc.1) ### How did you install the app? App Store, probably. I downloaded last summer and had no issues until now ### Homeserver Matrix ### Will you send logs? Yes
1.0
issue with keys - ### Steps to reproduce 1. Where are you starting? What can you see? Bisq General room 2. What do you click? Not clicking anything. Scanning messages in the room, I can see an apparent struggle to decrypt messages going on. 3. More steps… My MacBook Pro started struggling to decrypt in an encrypted room and despite signing in and out of Element, I continue to receive the following error message "Some messages could not be decrypted Unfortunately, there are no other verified devices to request decryption keys from. Signing in and verifying other devices may help avoid this situation in the future." I left the room and tried to reenter and was urged to open an issue here due to an error. ### Outcome #### What did you expect? #### What happened instead? error message keeps reappearing. Struggle to decrypt iMessages. Final error message recited above. Messages in room prior to January 24 no longer appear. ### Operating system Mac OS 10.15.7 ### Application version Version 1.11.21-rc.1 (1.11.21-rc.1) ### How did you install the app? App Store, probably. I downloaded last summer and had no issues until now ### Homeserver Matrix ### Will you send logs? Yes
defect
issue with keys steps to reproduce where are you starting what can you see bisq general room what do you click not clicking anything scanning messages in the room i can see an apparent struggle to decrypt messages going on more steps… my macbook pro started struggling to decrypt in an encrypted room and despite signing in and out of element i continue to receive the following error message some messages could not be decrypted unfortunately there are no other verified devices to request decryption keys from signing in and verifying other devices may help avoid this situation in the future i left the room and tried to reenter and was urged to open an issue here due to an error outcome what did you expect what happened instead error message keeps reappearing struggle to decrypt imessages final error message recited above messages in room prior to january no longer appear operating system mac os application version version rc rc how did you install the app app store probably i downloaded last summer and had no issues until now homeserver matrix will you send logs yes
1
543,434
15,881,942,773
IssuesEvent
2021-04-09 15:24:24
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Revision dropdown goes blank when creating a new version
API-M 4.0.0 Feature/Revisioning Priority/High React-UI Type/Bug
### Description: The revision select dropdown in the Top Menu bar in Publisher goes blank when creating a new API version. Current API option is not selected. It shows up after page reload. <img width="1653" alt="Screenshot 2021-04-02 at 17 22 48" src="https://user-images.githubusercontent.com/8557410/113413587-a1b2ca00-93d8-11eb-80d1-55553102fe67.png">
1.0
Revision dropdown goes blank when creating a new version - ### Description: The revision select dropdown in the Top Menu bar in Publisher goes blank when creating a new API version. Current API option is not selected. It shows up after page reload. <img width="1653" alt="Screenshot 2021-04-02 at 17 22 48" src="https://user-images.githubusercontent.com/8557410/113413587-a1b2ca00-93d8-11eb-80d1-55553102fe67.png">
non_defect
revision dropdown goes blank when creating a new version description the revision select dropdown in the top menu bar in publisher goes blank when creating a new api version current api option is not selected it shows up after page reload img width alt screenshot at src
0
326,615
28,006,793,738
IssuesEvent
2023-03-27 15:44:35
primitivefinance/arbiter
https://api.github.com/repos/primitivefinance/arbiter
closed
revm error handling
Needs More Information Testing & Error Handling
We want to be better about our error handling when interacting with revm. We encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until @Autoparallel added the following lines to propagate the error for us. https://github.com/primitivefinance/arbiter/blob/0e90313fc2caeb31839cdd0d4c4c5a10762563fb/crates/simulate/src/execution.rs#L95-L99
1.0
revm error handling - We want to be better about our error handling when interacting with revm. We encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until @Autoparallel added the following lines to propagate the error for us. https://github.com/primitivefinance/arbiter/blob/0e90313fc2caeb31839cdd0d4c4c5a10762563fb/crates/simulate/src/execution.rs#L95-L99
non_defect
revm error handling we want to be better about our error handling when interacting with revm we encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until autoparallel added the following lines to propagate the error for us
0
67,254
20,961,596,559
IssuesEvent
2022-03-27 21:46:23
abedmaatalla/imsdroid
https://api.github.com/repos/abedmaatalla/imsdroid
closed
sip register to server in android-ngn-stack
Priority-Medium Type-Defect auto-migrated
``` when you register to the server use a wrong user information, should get the `NgnRegistrationEventTypes.REGISTRATION_NOK` event, but actually receive the `NgnRegistrationEventTypes.UNREGISTRATION_OK` event. when i check the code in `NgnSipService` class, method `public int OnDialogEvent(DialogEvent e)` , switch code `case tinyWRAPConstants.tsip_event_code_dialog_terminated:`, you directed broadcast event `NgnRegistrationEventTypes.UNREGISTRATION_OK`, and in android-ngn-stack all code don't have `NgnRegistrationEventTypes.REGISTRATION_NOK` event broadcast. ``` Original issue reported on code.google.com by `idiottig...@gmail.com` on 8 Feb 2012 at 1:39
1.0
sip register to server in android-ngn-stack - ``` when you register to the server use a wrong user information, should get the `NgnRegistrationEventTypes.REGISTRATION_NOK` event, but actually receive the `NgnRegistrationEventTypes.UNREGISTRATION_OK` event. when i check the code in `NgnSipService` class, method `public int OnDialogEvent(DialogEvent e)` , switch code `case tinyWRAPConstants.tsip_event_code_dialog_terminated:`, you directed broadcast event `NgnRegistrationEventTypes.UNREGISTRATION_OK`, and in android-ngn-stack all code don't have `NgnRegistrationEventTypes.REGISTRATION_NOK` event broadcast. ``` Original issue reported on code.google.com by `idiottig...@gmail.com` on 8 Feb 2012 at 1:39
defect
sip register to server in android ngn stack when you register to the server use a wrong user information should get the ngnregistrationeventtypes registration nok event but actually receive the ngnregistrationeventtypes unregistration ok event when i check the code in ngnsipservice class method public int ondialogevent dialogevent e switch code case tinywrapconstants tsip event code dialog terminated you directed broadcast event ngnregistrationeventtypes unregistration ok and in android ngn stack all code don t have ngnregistrationeventtypes registration nok event broadcast original issue reported on code google com by idiottig gmail com on feb at
1
1,218
2,601,760,220
IssuesEvent
2015-02-24 00:34:48
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
Memory Leak "fix" will bug if you leave the game instead of restart it
auto-migrated Component-Logic Milestone-Release Priority-Critical Type-Defect Usability
``` To do. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 25 Jan 2011 at 11:14
1.0
Memory Leak "fix" will bug if you leave the game instead of restart it - ``` To do. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 25 Jan 2011 at 11:14
defect
memory leak fix will bug if you leave the game instead of restart it to do original issue reported on code google com by aheinerm on jan at
1
47,532
13,056,224,755
IssuesEvent
2020-07-30 04:02:44
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
I3File does not correctly follow Python iterator interface (Trac #688)
Migrated from Trac dataio defect
The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`. Reference: http://docs.python.org/library/stdtypes.html#iterator-types In rare cases, this issue can cause unexpected iterator behavior. For example: ```text 1 it = iter(i3file) 2 frame1 = it.next() 3 for frame in it: 4 # attempt to act on second and all subsequent frames... ``` On the first run through the loop, at line 4, `frame` will equal `frame1`. I'm giving this a low priority, but figured it should be documented. Migrated from https://code.icecube.wisc.edu/ticket/688 ```json { "status": "closed", "changetime": "2014-03-22T04:28:38", "description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.", "reporter": "sjackso", "cc": "", "resolution": "fixed", "_ts": "1395462518000000", "component": "dataio", "summary": "I3File does not correctly follow Python iterator interface", "priority": "minor", "keywords": "I3File I3SequentialFile iterator", "time": "2012-09-27T19:52:15", "milestone": "", "owner": "sjackso", "type": "defect" } ```
1.0
I3File does not correctly follow Python iterator interface (Trac #688) - The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`. Reference: http://docs.python.org/library/stdtypes.html#iterator-types In rare cases, this issue can cause unexpected iterator behavior. For example: ```text 1 it = iter(i3file) 2 frame1 = it.next() 3 for frame in it: 4 # attempt to act on second and all subsequent frames... ``` On the first run through the loop, at line 4, `frame` will equal `frame1`. I'm giving this a low priority, but figured it should be documented. Migrated from https://code.icecube.wisc.edu/ticket/688 ```json { "status": "closed", "changetime": "2014-03-22T04:28:38", "description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.", "reporter": "sjackso", "cc": "", "resolution": "fixed", "_ts": "1395462518000000", "component": "dataio", "summary": "I3File does not correctly follow Python iterator interface", "priority": "minor", "keywords": "I3File I3SequentialFile iterator", "time": "2012-09-27T19:52:15", "milestone": "", "owner": "sjackso", "type": "defect" } ```
defect
does not correctly follow python iterator interface trac the python binding for an implemented in c as provides methods next and iter to implement the python iterator interface however it does not support the iterator interface correctly because an is both a container and an iterator a correct implementation would provide a separate iterator class that implemented next and iter while itself would only provide iter reference in rare cases this issue can cause unexpected iterator behavior for example text it iter it next for frame in it attempt to act on second and all subsequent frames on the first run through the loop at line frame will equal i m giving this a low priority but figured it should be documented migrated from json status closed changetime description the python binding for an implemented in c as provides methods next and iter to implement the python iterator interface however it does not support the iterator interface correctly because an is both a container and an iterator a correct implementation would provide a separate iterator class that implemented next and iter while itself would only provide iter n nreference rare cases this issue can cause unexpected iterator behavior for example n n it iter it next for frame in it attempt to act on second and all subsequent frames n n non the first run through the loop at line frame will equal n ni m giving this a low priority but figured it should be documented reporter sjackso cc resolution fixed ts component dataio summary does not correctly follow python iterator interface priority minor keywords iterator time milestone owner sjackso type defect
1
238,775
26,155,044,117
IssuesEvent
2022-12-30 20:00:12
labsai/EDDI
https://api.github.com/repos/labsai/EDDI
closed
quarkus-smallrye-openapi-2.14.1.Final.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quarkus-smallrye-openapi-2.14.1.Final.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quarkus-smallrye-openapi version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary> ### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p> <p> Dependency Hierarchy: - quarkus-smallrye-openapi-2.14.1.Final.jar (Root Library) - smallrye-open-api-core-2.3.1.jar - jackson-dataformat-yaml-2.13.4.jar - :x: **snakeyaml-1.33.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p> <p>Release Date: 2022-12-01</p> <p>Fix Resolution: org.yaml:snakeyaml - 1.31</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
quarkus-smallrye-openapi-2.14.1.Final.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quarkus-smallrye-openapi-2.14.1.Final.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quarkus-smallrye-openapi version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary> ### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p> <p> Dependency Hierarchy: - quarkus-smallrye-openapi-2.14.1.Final.jar (Root Library) - smallrye-open-api-core-2.3.1.jar - jackson-dataformat-yaml-2.13.4.jar - :x: **snakeyaml-1.33.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p> <p>Release Date: 2022-12-01</p> <p>Fix Resolution: org.yaml:snakeyaml - 1.31</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_defect
quarkus smallrye openapi final jar vulnerabilities highest severity is autoclosed vulnerable library quarkus smallrye openapi final jar path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar vulnerabilities cve severity cvss dependency type fixed in quarkus smallrye openapi version remediation available high snakeyaml jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy quarkus smallrye openapi final jar root library smallrye open api core jar jackson dataformat yaml jar x snakeyaml jar vulnerable library found in base branch main vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
0
298,738
22,554,518,241
IssuesEvent
2022-06-27 09:03:20
dell/ansible-powerscale
https://api.github.com/repos/dell/ansible-powerscale
closed
Create Network pool documentation is wrong
documentation
@rajendravarma077 @meshuga @anupamaloke @walker2 @dattaarindam From Documentation, below is the code to create a network pool. But getting below error stating unsupported parameters name fatal: [localhost]: FAILED! => {"changed": false, "msg": "Unsupported parameters for (dellemc.powerscale.dellemc_powerscale_subnet) module: access_zone, pool Supported parameters include: api_password, api_user, description, gateway_priority, groupnet_name, netmask, new_subnet_name, onefs_host, port_no, state, subnet_name, subnet_params, verify_ssl"} Do we need to use dellemc_powerscale_subnet or dellemc_powerscale_networkpool. - name: Create Network Pool dellemc_powerscale_subnet: onefs_host: "{{onefs_host}}" api_user: "{{api_user}}" api_password: "{{api_password}}" verify_ssl: "{{verify_ssl}}" groupnet: "groupnet0" subnet: "subnet0" pool: "Test_Pool_2" access_zone: "system" state: "present"
1.0
Create Network pool documentation is wrong - @rajendravarma077 @meshuga @anupamaloke @walker2 @dattaarindam From Documentation, below is the code to create a network pool. But getting below error stating unsupported parameters name fatal: [localhost]: FAILED! => {"changed": false, "msg": "Unsupported parameters for (dellemc.powerscale.dellemc_powerscale_subnet) module: access_zone, pool Supported parameters include: api_password, api_user, description, gateway_priority, groupnet_name, netmask, new_subnet_name, onefs_host, port_no, state, subnet_name, subnet_params, verify_ssl"} Do we need to use dellemc_powerscale_subnet or dellemc_powerscale_networkpool. - name: Create Network Pool dellemc_powerscale_subnet: onefs_host: "{{onefs_host}}" api_user: "{{api_user}}" api_password: "{{api_password}}" verify_ssl: "{{verify_ssl}}" groupnet: "groupnet0" subnet: "subnet0" pool: "Test_Pool_2" access_zone: "system" state: "present"
non_defect
create network pool documentation is wrong meshuga anupamaloke dattaarindam from documentation below is the code to create a network pool but getting below error stating unsupported parameters name fatal failed changed false msg unsupported parameters for dellemc powerscale dellemc powerscale subnet module access zone pool supported parameters include api password api user description gateway priority groupnet name netmask new subnet name onefs host port no state subnet name subnet params verify ssl do we need to use dellemc powerscale subnet or dellemc powerscale networkpool name create network pool dellemc powerscale subnet onefs host onefs host api user api user api password api password verify ssl verify ssl groupnet subnet pool test pool access zone system state present
0
71,912
23,850,012,401
IssuesEvent
2022-09-06 17:00:48
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Markdown doesn't work in threaded replies
T-Defect A-Markdown A-Timeline S-Major O-Occasional
### Steps to reproduce 1. See a threaded message with formatting, e.g. "**message**" 2. Reply to it ### Outcome #### What did you expect? > **message** #### What happened instead? > \*\*message\*\* ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store 1.4.34 ### Homeserver matrix.org ### Will you send logs? No ### Are you willing to provide a PR? No
1.0
Markdown doesn't work in threaded replies - ### Steps to reproduce 1. See a threaded message with formatting, e.g. "**message**" 2. Reply to it ### Outcome #### What did you expect? > **message** #### What happened instead? > \*\*message\*\* ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store 1.4.34 ### Homeserver matrix.org ### Will you send logs? No ### Are you willing to provide a PR? No
defect
markdown doesn t work in threaded replies steps to reproduce see a threaded message with formatting e g message reply to it outcome what did you expect message what happened instead message your phone model no response operating system version no response application version and app store homeserver matrix org will you send logs no are you willing to provide a pr no
1
236,298
19,530,561,947
IssuesEvent
2021-12-30 15:59:53
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks - useMetricsExplorerData Hook should paginate
failed-test Team:Infra Monitoring UI
A test failed on a tracked branch ``` Error: Timed out in waitForNextUpdate after 1000ms. at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13) at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks/use_metrics_explorer_data.test.tsx:96:5) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [CI Build - 7.x](https://buildkite.com/elastic/kibana-hourly/builds/1617#9b2d480c-5a37-4a17-9a89-2fd88b3fdfbf) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks","test.name":"useMetricsExplorerData Hook should paginate","test.failCount":3}} -->
1.0
Failing test: Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks - useMetricsExplorerData Hook should paginate - A test failed on a tracked branch ``` Error: Timed out in waitForNextUpdate after 1000ms. at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13) at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks/use_metrics_explorer_data.test.tsx:96:5) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [CI Build - 7.x](https://buildkite.com/elastic/kibana-hourly/builds/1617#9b2d480c-5a37-4a17-9a89-2fd88b3fdfbf) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks","test.name":"useMetricsExplorerData Hook should paginate","test.failCount":3}} -->
non_defect
failing test jest tests x pack plugins infra public pages metrics metrics explorer hooks usemetricsexplorerdata hook should paginate a test failed on a tracked branch error timed out in waitfornextupdate after at waitfornextupdate opt local ssd buildkite builds kb elastic kibana hourly kibana node modules testing library react hooks lib core asyncutils js at object opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins infra public pages metrics metrics explorer hooks use metrics explorer data test tsx at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js first failure
0
62,331
17,023,899,382
IssuesEvent
2021-07-03 04:26:47
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Bounding box formated as double for a result in jsonv2 format
Component: nominatim Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 12.28pm, Sunday, 2nd March 2014]** When I was searching for san francisco using GNOME Maps I got a crash in the geocode-glib library. It could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings. You can see the result here: http://nominatim.openstreetmap.org/search.php?q=San%20Francisco,%20Agusan%20Del%20Sur,%20Republic%20of%20the%20Philippines&format=json&addressdetails=1&limit=11 Note that the second place has the bounding box as double and not string.
1.0
Bounding box formated as double for a result in jsonv2 format - **[Submitted to the original trac issue database at 12.28pm, Sunday, 2nd March 2014]** When I was searching for san francisco using GNOME Maps I got a crash in the geocode-glib library. It could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings. You can see the result here: http://nominatim.openstreetmap.org/search.php?q=San%20Francisco,%20Agusan%20Del%20Sur,%20Republic%20of%20the%20Philippines&format=json&addressdetails=1&limit=11 Note that the second place has the bounding box as double and not string.
defect
bounding box formated as double for a result in format when i was searching for san francisco using gnome maps i got a crash in the geocode glib library it could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings you can see the result here note that the second place has the bounding box as double and not string
1
36,722
8,093,388,716
IssuesEvent
2018-08-10 00:31:50
CenturyLinkCloud/mdw
https://api.github.com/repos/CenturyLinkCloud/mdw
closed
CLI Import deletes non-mdw local assets
defect
When running `mdw update` to import mdw base assets, all local assets beginning with "com/" are deleted.
1.0
CLI Import deletes non-mdw local assets - When running `mdw update` to import mdw base assets, all local assets beginning with "com/" are deleted.
defect
cli import deletes non mdw local assets when running mdw update to import mdw base assets all local assets beginning with com are deleted
1
43,553
11,747,012,810
IssuesEvent
2020-03-12 12:53:44
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
Events listing pagination and sorting is broken
Defect Outreach hub VAMC system
**Describe the defect** Events pagination is whacked Event sorting is all over the place for this one https://staging.va.gov/outreach-and-events/events/ it seems to be date ASC on each page, but not across all 3 pages. Three pages exist here, but there should only be one: https://staging.va.gov/pittsburgh-health-care/events/ * A past featured event shows on all pages. * Pages 2 and 3 are empty except for the featured event **To Reproduce** Steps to reproduce the behavior: 1. Go to above URLs and click around **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** ![VA_Pittsburgh_Health_Care___Events___Veterans_Affairs.jpg](https://images.zenhubusercontent.com/5d6e57ad330f64311395f08b/50a6a842-c79b-484e-82af-dd8b0e6b47d6) **Desktop (please complete the following information if relevant, or delete):** - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here. ## Labels - [/] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [/] CMS objective (orange) (not needed for bug tickets) - [ ] CMS-supported product (black)
1.0
Events listing pagination and sorting is broken - **Describe the defect** Events pagination is whacked Event sorting is all over the place for this one https://staging.va.gov/outreach-and-events/events/ it seems to be date ASC on each page, but not across all 3 pages. Three pages exist here, but there should only be one: https://staging.va.gov/pittsburgh-health-care/events/ * A past featured event shows on all pages. * Pages 2 and 3 are empty except for the featured event **To Reproduce** Steps to reproduce the behavior: 1. Go to above URLs and click around **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** ![VA_Pittsburgh_Health_Care___Events___Veterans_Affairs.jpg](https://images.zenhubusercontent.com/5d6e57ad330f64311395f08b/50a6a842-c79b-484e-82af-dd8b0e6b47d6) **Desktop (please complete the following information if relevant, or delete):** - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here. ## Labels - [/] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [/] CMS objective (orange) (not needed for bug tickets) - [ ] CMS-supported product (black)
defect
events listing pagination and sorting is broken describe the defect events pagination is whacked event sorting is all over the place for this one it seems to be date asc on each page but not across all pages three pages exist here but there should only be one a past featured event shows on all pages pages and are empty except for the featured event to reproduce steps to reproduce the behavior go to above urls and click around expected behavior a clear and concise description of what you expected to happen screenshots desktop please complete the following information if relevant or delete os browser version additional context add any other context about the problem here labels issue type red defaults to defect cms subsystem green cms practice area blue cms objective orange not needed for bug tickets cms supported product black
1
287,956
21,677,925,236
IssuesEvent
2022-05-09 00:54:17
hfang821/u-develop-it-BE
https://api.github.com/repos/hfang821/u-develop-it-BE
closed
Create the voters table
documentation
**User Story** * As a user, I can request a list of voters alphabetized by last name. * As a user, I can request a single voter's information. * As a user, I can create a voter. * As a user, I can update a voter's email. * As a user, I can delete a voter
1.0
Create the voters table - **User Story** * As a user, I can request a list of voters alphabetized by last name. * As a user, I can request a single voter's information. * As a user, I can create a voter. * As a user, I can update a voter's email. * As a user, I can delete a voter
non_defect
create the voters table user story as a user i can request a list of voters alphabetized by last name as a user i can request a single voter s information as a user i can create a voter as a user i can update a voter s email as a user i can delete a voter
0
41,523
10,508,465,011
IssuesEvent
2019-09-27 08:43:11
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
LikeEscapeStep#escape() also supported by Teradata
C: DB: Teradata C: DB: Teradata Aster E: Enterprise Edition P: Medium R: Fixed T: Defect
The `@Support` annotation on `LikeEscapeStep#escape()` lists most dialects, but `TERADATA` appears to be missing. According to the official documentation the `ESCAPE` clause is supported: https://docs.teradata.com/reader/756LNiPSFdY~4JcCCcR5Cw/ZP3CE_cR~e7V50zVkzzeVQ.
1.0
LikeEscapeStep#escape() also supported by Teradata - The `@Support` annotation on `LikeEscapeStep#escape()` lists most dialects, but `TERADATA` appears to be missing. According to the official documentation the `ESCAPE` clause is supported: https://docs.teradata.com/reader/756LNiPSFdY~4JcCCcR5Cw/ZP3CE_cR~e7V50zVkzzeVQ.
defect
likeescapestep escape also supported by teradata the support annotation on likeescapestep escape lists most dialects but teradata appears to be missing according to the official documentation the escape clause is supported
1
20,081
3,295,315,199
IssuesEvent
2015-10-31 20:48:32
chief-atx/bcmon
https://api.github.com/repos/chief-atx/bcmon
closed
No development since April 2013
auto-migrated Priority-Medium Type-Defect
``` I am forking this project. Is anyone on the planet still interested in maintaining and/or contributing? ``` Original issue reported on code.google.com by `nethers...@gmail.com` on 20 Mar 2015 at 6:07
1.0
No development since April 2013 - ``` I am forking this project. Is anyone on the planet still interested in maintaining and/or contributing? ``` Original issue reported on code.google.com by `nethers...@gmail.com` on 20 Mar 2015 at 6:07
defect
no development since april i am forking this project is anyone on the planet still interested in maintaining and or contributing original issue reported on code google com by nethers gmail com on mar at
1
390,407
11,543,316,877
IssuesEvent
2020-02-18 09:23:10
cthit/Gamma
https://api.github.com/repos/cthit/Gamma
opened
Cannot add group post
Priority: Critical Status: Accepted Where: Backend
https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/controller/admin/GroupPostAdminController.java#L61 Tries to see if the swedish name exists, but `postExists` expects an UUID. https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/service/PostService.java#L44
1.0
Cannot add group post - https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/controller/admin/GroupPostAdminController.java#L61 Tries to see if the swedish name exists, but `postExists` expects an UUID. https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/service/PostService.java#L44
non_defect
cannot add group post tries to see if the swedish name exists but postexists expects an uuid
0
53,493
13,261,753,733
IssuesEvent
2020-08-20 20:28:23
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[topsimulator] has a number of #warnings (Trac #1529)
Migrated from Trac combo simulation defect
These look like TODOs <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1529">https://code.icecube.wisc.edu/projects/icecube/ticket/1529</a>, reported by kjmeagherand owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:35", "_ts": "1550067215093672", "description": "These look like TODOs", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-01-26T10:20:59", "component": "combo simulation", "summary": "[topsimulator] has a number of #warnings", "priority": "normal", "keywords": "", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
1.0
[topsimulator] has a number of #warnings (Trac #1529) - These look like TODOs <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1529">https://code.icecube.wisc.edu/projects/icecube/ticket/1529</a>, reported by kjmeagherand owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:35", "_ts": "1550067215093672", "description": "These look like TODOs", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-01-26T10:20:59", "component": "combo simulation", "summary": "[topsimulator] has a number of #warnings", "priority": "normal", "keywords": "", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
defect
has a number of warnings trac these look like todos migrated from json status closed changetime ts description these look like todos reporter kjmeagher cc resolution fixed time component combo simulation summary has a number of warnings priority normal keywords milestone owner jgonzalez type defect
1
5,238
2,610,183,905
IssuesEvent
2015-02-26 18:58:30
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
证实长了色斑该怎么办
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。长了色斑该怎么办, 《客户案例》   因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比�� �了解,现在我要说的是还是精华祛斑比较好,我就是用精华� ��斑的方法祛斑的。<br>   我是因为工作的关系需要经常化妆,大家都知道,化妆�� �都会或多或少的含有一些化学成分,慢慢的,我的脸上就长� ��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道 斑也分很多种,有先天的,也有后天的,先天的一般是很难�� �除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天� ��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法 完全去掉的。刚开始也是受化妆品的影响,用了很多的美白�� �斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又� ��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知 识,才知道那些快速口服药中一般含有重金属,短期效果明�� �,但是一旦停用,就会反复长更多的斑,而且这样的情况是� ��难被改善的,所以我就再不用那些美白祛斑的口服的了。<br >   我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个�� �纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯� ��华的应该没什么副作用,再说是朋友介绍的,应该是有用, 就去他们商城上详细询问了客服,告诉我这个主要是通过调�� �祛斑的,对身体没任何副作用,我又查了这个产品的各个成� ��的作用,确实还挺不错的,就一下订购了三个周期的,真没 想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,�� �在皮肤也好了很多,也没反弹。 阅读了长了色斑该怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 长了色斑该怎么办,同时为您分享祛斑小方法 天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙�� �光泽 必须长期内调,做好补血养气的工作才能让自己摆脱天生长�� �的问题。红枣、阿胶、 红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物�� �很好的补气作用。当然除了补气补血外,体内调理好了,那� ��长斑的机会就更少了。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:11
1.0
证实长了色斑该怎么办 - ``` 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。长了色斑该怎么办, 《客户案例》   因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比�� �了解,现在我要说的是还是精华祛斑比较好,我就是用精华� ��斑的方法祛斑的。<br>   我是因为工作的关系需要经常化妆,大家都知道,化妆�� �都会或多或少的含有一些化学成分,慢慢的,我的脸上就长� ��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道 斑也分很多种,有先天的,也有后天的,先天的一般是很难�� �除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天� ��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法 完全去掉的。刚开始也是受化妆品的影响,用了很多的美白�� �斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又� ��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知 识,才知道那些快速口服药中一般含有重金属,短期效果明�� �,但是一旦停用,就会反复长更多的斑,而且这样的情况是� ��难被改善的,所以我就再不用那些美白祛斑的口服的了。<br >   我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个�� �纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯� ��华的应该没什么副作用,再说是朋友介绍的,应该是有用, 就去他们商城上详细询问了客服,告诉我这个主要是通过调�� �祛斑的,对身体没任何副作用,我又查了这个产品的各个成� ��的作用,确实还挺不错的,就一下订购了三个周期的,真没 想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,�� �在皮肤也好了很多,也没反弹。 阅读了长了色斑该怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 长了色斑该怎么办,同时为您分享祛斑小方法 天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙�� �光泽 必须长期内调,做好补血养气的工作才能让自己摆脱天生长�� �的问题。红枣、阿胶、 红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物�� �很好的补气作用。当然除了补气补血外,体内调理好了,那� ��长斑的机会就更少了。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:11
defect
证实长了色斑该怎么办 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。长了色斑该怎么办, 《客户案例》   因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比�� �了解,现在我要说的是还是精华祛斑比较好,我就是用精华� ��斑的方法祛斑的。   我是因为工作的关系需要经常化妆,大家都知道,化妆�� �都会或多或少的含有一些化学成分,慢慢的,我的脸上就长� ��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道 斑也分很多种,有先天的,也有后天的,先天的一般是很难�� �除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天� ��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法 完全去掉的。刚开始也是受化妆品的影响,用了很多的美白�� �斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又� ��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知 识,才知道那些快速口服药中一般含有重金属,短期效果明�� �,但是一旦停用,就会反复长更多的斑,而且这样的情况是� ��难被改善的,所以我就再不用那些美白祛斑的口服的了。 br   我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个�� �纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯� ��华的应该没什么副作用,再说是朋友介绍的,应该是有用, 就去他们商城上详细询问了客服,告诉我这个主要是通过调�� �祛斑的,对身体没任何副作用,我又查了这个产品的各个成� ��的作用,确实还挺不错的,就一下订购了三个周期的,真没 想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,�� �在皮肤也好了很多,也没反弹。 阅读了长了色斑该怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 长了色斑该怎么办,同时为您分享祛斑小方法 天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙�� �光泽 必须长期内调,做好补血养气的工作才能让自己摆脱天生长�� �的问题。红枣、阿胶、 红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物�� �很好的补气作用。当然除了补气补血外,体内调理好了,那� ��长斑的机会就更少了。 original issue reported on code google com by additive gmail com on jul at
1
68,774
21,886,101,546
IssuesEvent
2022-05-19 18:48:33
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Excess parentheses generated in unary ~ (QOM.BitNot) or - (QOM.Neg) operators
T: Defect C: Functionality P: Medium E: All Editions
Like https://github.com/jOOQ/jOOQ/issues/7362, we tend to produce too many parentheses around the argument of unary `~` (`QOM.BitNot`) and `-` (`QOM.Neg`) operators. E.g. Input: ```sql select ~(1 + 2), -(1 + 2) ``` Output: ```sql select ~((1 + 2)), -((1 + 2)) ```
1.0
Excess parentheses generated in unary ~ (QOM.BitNot) or - (QOM.Neg) operators - Like https://github.com/jOOQ/jOOQ/issues/7362, we tend to produce too many parentheses around the argument of unary `~` (`QOM.BitNot`) and `-` (`QOM.Neg`) operators. E.g. Input: ```sql select ~(1 + 2), -(1 + 2) ``` Output: ```sql select ~((1 + 2)), -((1 + 2)) ```
defect
excess parentheses generated in unary qom bitnot or qom neg operators like we tend to produce too many parentheses around the argument of unary qom bitnot and qom neg operators e g input sql select output sql select
1
10,850
2,622,193,255
IssuesEvent
2015-03-04 00:24:01
byzhang/cudpp
https://api.github.com/repos/byzhang/cudpp
opened
make install omits cudpp_config.h
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. mkdir -p build; cd build; 2. cmake .. 3. make 4. sudo make install What is the expected output? What do you see instead? Expect /usr/local/include to contain cudpp_config.h, but it isn't there. This results in the following error when building a 3rd-party project that depends on cudpp: /usr/local/include/cudpp.h:501:26: fatal error: cudpp_config.h: No such file or directory What version of the product are you using? On what operating system? cudpp 2.0 ubuntu 12.04 cmake 2.8.7 Please provide any additional information below. ``` Original issue reported on code.google.com by `ksi...@email.arizona.edu` on 5 Jun 2012 at 11:12
1.0
make install omits cudpp_config.h - ``` What steps will reproduce the problem? 1. mkdir -p build; cd build; 2. cmake .. 3. make 4. sudo make install What is the expected output? What do you see instead? Expect /usr/local/include to contain cudpp_config.h, but it isn't there. This results in the following error when building a 3rd-party project that depends on cudpp: /usr/local/include/cudpp.h:501:26: fatal error: cudpp_config.h: No such file or directory What version of the product are you using? On what operating system? cudpp 2.0 ubuntu 12.04 cmake 2.8.7 Please provide any additional information below. ``` Original issue reported on code.google.com by `ksi...@email.arizona.edu` on 5 Jun 2012 at 11:12
defect
make install omits cudpp config h what steps will reproduce the problem mkdir p build cd build cmake make sudo make install what is the expected output what do you see instead expect usr local include to contain cudpp config h but it isn t there this results in the following error when building a party project that depends on cudpp usr local include cudpp h fatal error cudpp config h no such file or directory what version of the product are you using on what operating system cudpp ubuntu cmake please provide any additional information below original issue reported on code google com by ksi email arizona edu on jun at
1
65,226
19,277,372,001
IssuesEvent
2021-12-10 13:28:22
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Polls appear as empty events when I don't have the Polls lab flag enabled
T-Defect
### Steps to reproduce 1. Disable Polls in labs 2. As another user with Polls enabled, create a poll ### Outcome #### What did you expect? I should see nothing #### What happened instead? I see an empty event ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version Element version: ebc2bb7c5644-react-69c847b94c7f-js-dd8c157bb95a Olm version: 3.2.3 ### Homeserver _No response_ ### Will you send logs? No
1.0
Polls appear as empty events when I don't have the Polls lab flag enabled - ### Steps to reproduce 1. Disable Polls in labs 2. As another user with Polls enabled, create a poll ### Outcome #### What did you expect? I should see nothing #### What happened instead? I see an empty event ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version Element version: ebc2bb7c5644-react-69c847b94c7f-js-dd8c157bb95a Olm version: 3.2.3 ### Homeserver _No response_ ### Will you send logs? No
defect
polls appear as empty events when i don t have the polls lab flag enabled steps to reproduce disable polls in labs as another user with polls enabled create a poll outcome what did you expect i should see nothing what happened instead i see an empty event operating system no response browser information no response url for webapp no response application version element version react js olm version homeserver no response will you send logs no
1
18,395
3,054,484,972
IssuesEvent
2015-08-13 03:06:56
eczarny/spectacle
https://api.github.com/repos/eczarny/spectacle
closed
AutomaticUpdateCheckEnabled not effective
defect ★
We deploy Spectacle to our fleet of school Macs with OS X 10.9 via munki. Most devices were at Spectacle 0.8.6. We manage all updates via munki; we never want any software to prompts to update. Our post-install-script includes the following: `# disable Spectacle autoupdate defaults write /Library/Preferences/com.divisiblebyzero.Spectacle AutomaticUpdateCheckEnabled -bool false` However, this morning with the release of 0.8.8, end-users are being prompted to upgrade. `defaults read /Library/Preferences/com.divisiblebyzero.Spectacle` on a non-upgraded machine (still running 0.8.6) yields, both before and after 0.8.8 is installed via munki: `{` ` AutomaticUpdateCheckEnabled = 0;` `}` Has the "disabling Automatic Update" mechanism changed? Thanks, Mike
1.0
AutomaticUpdateCheckEnabled not effective - We deploy Spectacle to our fleet of school Macs with OS X 10.9 via munki. Most devices were at Spectacle 0.8.6. We manage all updates via munki; we never want any software to prompts to update. Our post-install-script includes the following: `# disable Spectacle autoupdate defaults write /Library/Preferences/com.divisiblebyzero.Spectacle AutomaticUpdateCheckEnabled -bool false` However, this morning with the release of 0.8.8, end-users are being prompted to upgrade. `defaults read /Library/Preferences/com.divisiblebyzero.Spectacle` on a non-upgraded machine (still running 0.8.6) yields, both before and after 0.8.8 is installed via munki: `{` ` AutomaticUpdateCheckEnabled = 0;` `}` Has the "disabling Automatic Update" mechanism changed? Thanks, Mike
defect
automaticupdatecheckenabled not effective we deploy spectacle to our fleet of school macs with os x via munki most devices were at spectacle we manage all updates via munki we never want any software to prompts to update our post install script includes the following disable spectacle autoupdate defaults write library preferences com divisiblebyzero spectacle automaticupdatecheckenabled bool false however this morning with the release of end users are being prompted to upgrade defaults read library preferences com divisiblebyzero spectacle on a non upgraded machine still running yields both before and after is installed via munki automaticupdatecheckenabled has the disabling automatic update mechanism changed thanks mike
1
12,333
2,692,107,692
IssuesEvent
2015-04-01 03:53:58
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
opened
Wrong case on filename if outputBy is "NamespacePath"
defect
#### bridge.json ``` { output: "Bridge\\output", outputBy: "NamespacePath" } ``` #### Demo.cs ``` using Bridge; namespace Demo.Utils { public class App { public string Test { get; set; } } } ``` `demo > utils.js` is expected, but `demo > Utils.js` is created. ![screenshot 2015-03-31 21 49 56](https://cloud.githubusercontent.com/assets/62210/6934323/e4f5d4fc-d7ef-11e4-9be5-f5b19b41dfb5.png)
1.0
Wrong case on filename if outputBy is "NamespacePath" - #### bridge.json ``` { output: "Bridge\\output", outputBy: "NamespacePath" } ``` #### Demo.cs ``` using Bridge; namespace Demo.Utils { public class App { public string Test { get; set; } } } ``` `demo > utils.js` is expected, but `demo > Utils.js` is created. ![screenshot 2015-03-31 21 49 56](https://cloud.githubusercontent.com/assets/62210/6934323/e4f5d4fc-d7ef-11e4-9be5-f5b19b41dfb5.png)
defect
wrong case on filename if outputby is namespacepath bridge json output bridge output outputby namespacepath demo cs using bridge namespace demo utils public class app public string test get set demo utils js is expected but demo utils js is created
1
13,464
2,759,260,009
IssuesEvent
2015-04-28 01:44:14
jsnjycwyz/gogo-tester
https://api.github.com/repos/jsnjycwyz/gogo-tester
closed
扫描要的时间越来越久了
auto-migrated Priority-Medium Type-Defect
``` 以前一下就扫除几十个出来,现在需要的时间越来越久了,�� �一个小时了,才扫出十个 ``` Original issue reported on code.google.com by `Cloud...@gmail.com` on 20 Apr 2015 at 3:03
1.0
扫描要的时间越来越久了 - ``` 以前一下就扫除几十个出来,现在需要的时间越来越久了,�� �一个小时了,才扫出十个 ``` Original issue reported on code.google.com by `Cloud...@gmail.com` on 20 Apr 2015 at 3:03
defect
扫描要的时间越来越久了 以前一下就扫除几十个出来,现在需要的时间越来越久了,�� �一个小时了,才扫出十个 original issue reported on code google com by cloud gmail com on apr at
1