Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
39,673 | 9,608,808,635 | IssuesEvent | 2019-05-12 10:00:29 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | [Android] Collapase menu button is replaced with Back button | C: Android - Manager E: 1 day P: Minor R: fixed T: Defect | **Describe the bug**
Top left menu button is replaced with a back button that is from Event log
Is this intentional?
**Steps To Reproduce**
1. Use the beta
2. Immediately see the behaviour
**Expected behavior**
"Hamburger" or collapse menu button should be still there...
**Screenshots**
The beta 7.14

7.4.53


**System Information (please complete the following information):**
- OS: Android 8.1 | 1.0 | [Android] Collapase menu button is replaced with Back button - **Describe the bug**
Top left menu button is replaced with a back button that is from Event log
Is this intentional?
**Steps To Reproduce**
1. Use the beta
2. Immediately see the behaviour
**Expected behavior**
"Hamburger" or collapse menu button should be still there...
**Screenshots**
The beta 7.14

7.4.53


**System Information (please complete the following information):**
- OS: Android 8.1 | defect | collapase menu button is replaced with back button describe the bug top left menu button is replaced with a back button that is from event log is this intentional steps to reproduce use the beta immediately see the behaviour expected behavior hamburger or collapse menu button should be still there screenshots the beta system information please complete the following information os android | 1 |
4,519 | 5,133,146,702 | IssuesEvent | 2017-01-11 02:01:56 | archco/cosmos-css | https://api.github.com/repos/archco/cosmos-css | opened | Variables 정리 | css enhancement infrastructure | #78 처리중에 알게된 sass의 색 변수에 대한 정리 http://sass-lang.com/styleguide/color 를 참고삼아
Theme와 관련된 변수들을 정리하자.
### Theme Colors List
* primary-color
* secondary-color
* text-color
* text-color-invert
* text-color-weak
* text-color-strong
* body-bg-color
* body-bg-color-shade
* body-bg-color-invert
* shadow-color
* link-color
* link-color-hover
* link-color-visited
* border-color | 1.0 | Variables 정리 - #78 처리중에 알게된 sass의 색 변수에 대한 정리 http://sass-lang.com/styleguide/color 를 참고삼아
Theme와 관련된 변수들을 정리하자.
### Theme Colors List
* primary-color
* secondary-color
* text-color
* text-color-invert
* text-color-weak
* text-color-strong
* body-bg-color
* body-bg-color-shade
* body-bg-color-invert
* shadow-color
* link-color
* link-color-hover
* link-color-visited
* border-color | non_defect | variables 정리 처리중에 알게된 sass의 색 변수에 대한 정리 를 참고삼아 theme와 관련된 변수들을 정리하자 theme colors list primary color secondary color text color text color invert text color weak text color strong body bg color body bg color shade body bg color invert shadow color link color link color hover link color visited border color | 0 |
217,324 | 24,329,026,992 | IssuesEvent | 2022-09-30 17:29:58 | hapifhir/hapi-fhir | https://api.github.com/repos/hapifhir/hapi-fhir | closed | CVE-2022-40154 (High) detected in woodstox-core-6.3.1.jar - autoclosed | security vulnerability | ## CVE-2022-40154 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.3.1.jar</b></p></summary>
<p>Woodstox is a high-performance XML processor that implements Stax (JSR-173),
SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /hapi-fhir-structures-dstu3/pom.xml</p>
<p>Path to vulnerable library: /repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **woodstox-core-6.3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430">ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to serialise XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40154>CVE-2022-40154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-40154 (High) detected in woodstox-core-6.3.1.jar - autoclosed - ## CVE-2022-40154 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.3.1.jar</b></p></summary>
<p>Woodstox is a high-performance XML processor that implements Stax (JSR-173),
SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /hapi-fhir-structures-dstu3/pom.xml</p>
<p>Path to vulnerable library: /repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar,/repository/com/fasterxml/woodstox/woodstox-core/6.3.1/woodstox-core-6.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **woodstox-core-6.3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hapifhir/hapi-fhir/commit/ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430">ee1cb4e3927a600d81a29fdad0bd3fd11d1f5430</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to serialise XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40154>CVE-2022-40154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in woodstox core jar autoclosed cve high severity vulnerability vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file hapi fhir structures pom xml path to vulnerable library repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy x woodstox core jar vulnerable library found in head commit a href found in base branch master vulnerability details those using xstream to serialise xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stack overflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
16,025 | 2,870,252,244 | IssuesEvent | 2015-06-07 00:37:28 | pdelia/away3d | https://api.github.com/repos/pdelia/away3d | closed | Two view.render() calls required to update geometry correctly | auto-migrated Priority-Medium Type-Defect | #79 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:38Z
```
What steps will reproduce the problem?
1. Compile and run the attached FlashDevelop project for FP10, Away3D trunk
(issue occurs in rev 2028 at least)
2. Press '>' to use two render calls. Press '<' to use a single render
call. Press Space to swap cameras
What is the expected output? What do you see instead?
The 3D model consists of a camera and a sphere in an ObjectContainer3D. The
camera is pointing directly at the sphere and so the sphere should be
rendered in the centre of the frame. When the ObjectContainer3D is rotated
and the view rendered the sphere appears offset from its correct location.
This problem can be worked around by calling view.render() twice in
succession.
```
Original issue reported on code.google.com by `brucewei...@googlemail.com` on 16 Nov 2009 at 6:43
Attachments:
* [ContainerProblem.zip](https://storage.googleapis.com/google-code-attachments/away3d/issue-79/comment-0/ContainerProblem.zip)
| 1.0 | Two view.render() calls required to update geometry correctly - #79 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:38Z
```
What steps will reproduce the problem?
1. Compile and run the attached FlashDevelop project for FP10, Away3D trunk
(issue occurs in rev 2028 at least)
2. Press '>' to use two render calls. Press '<' to use a single render
call. Press Space to swap cameras
What is the expected output? What do you see instead?
The 3D model consists of a camera and a sphere in an ObjectContainer3D. The
camera is pointing directly at the sphere and so the sphere should be
rendered in the centre of the frame. When the ObjectContainer3D is rotated
and the view rendered the sphere appears offset from its correct location.
This problem can be worked around by calling view.render() twice in
succession.
```
Original issue reported on code.google.com by `brucewei...@googlemail.com` on 16 Nov 2009 at 6:43
Attachments:
* [ContainerProblem.zip](https://storage.googleapis.com/google-code-attachments/away3d/issue-79/comment-0/ContainerProblem.zip)
| defect | two view render calls required to update geometry correctly issue by googlecodeexporter created on what steps will reproduce the problem compile and run the attached flashdevelop project for trunk issue occurs in rev at least press to use two render calls press to use a single render call press space to swap cameras what is the expected output what do you see instead the model consists of a camera and a sphere in an the camera is pointing directly at the sphere and so the sphere should be rendered in the centre of the frame when the is rotated and the view rendered the sphere appears offset from its correct location this problem can be worked around by calling view render twice in succession original issue reported on code google com by brucewei googlemail com on nov at attachments | 1 |
23,390 | 3,807,888,227 | IssuesEvent | 2016-03-25 11:39:20 | PowerDNS/pdns | https://api.github.com/repos/PowerDNS/pdns | closed | pdns crashes when database insert fails to postgres database (lowercase constraint) | auth defect | In the current pdns setup, when the AXFR transfer contains names which has uppercase letters the following happens:
```
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Transaction started for 'nnn.fi.'
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Unable to feed record during incoming AXFR of 'nnn.fi.': GSQLBackend unable to feed record: Fatal error during query: insert into records (content,ttl,prio,type,domain_id,disabled,name,ordername,auth,change_date) values ($1,$2,$3,$4,$5,$6,$7,$8,$9,null): ERROR: new row for relation "records" violates check constraint "c_lowercase_name"
Mar 16 04:58:37 ns3 pdns_server: DETAIL: Failing row contains (6802, 5, 0A.client.nnn.fi, A, 192.168.0.10, 300, 0, null, f, null, t).
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Aborting possible open transaction for domain 'nnn.fi.' AXFR
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Got a signal 11, attempting to print trace:
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0x18195b) [0x7fde2958495b]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(+0x35670) [0x7fde27948670]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/pgsql-9.5/lib/libpq.so.5(PQclear+0x69) [0x7fde272c0f49]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN15SPgSQLStatementD0Ev+0x5d) [0x7fde274f4f5d]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN11GSQLBackend14freeStatementsEv+0x199) [0x7fde294967d9]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN13gPgSQLBackendD0Ev+0x34) [0x7fde274f1d24]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackend7cleanupEv+0x184) [0x7fde295d9f84]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackendD2Ev+0x2b) [0x7fde295da03b]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass4suckERK7DNSNameRKSs+0x5fc) [0x7fde295b211c]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass19retrievalLoopThreadEv+0x136) [0x7fde294aded6]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0xaea69) [0x7fde294b1a69]
Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libpthread.so.0(+0x7dc5) [0x7fde27cdbdc5]
Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(clone+0x6d) [0x7fde27a0928d]
``` | 1.0 | pdns crashes when database insert fails to postgres database (lowercase constraint) - In the current pdns setup, when the AXFR transfer contains names which has uppercase letters the following happens:
```
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Transaction started for 'nnn.fi.'
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Unable to feed record during incoming AXFR of 'nnn.fi.': GSQLBackend unable to feed record: Fatal error during query: insert into records (content,ttl,prio,type,domain_id,disabled,name,ordername,auth,change_date) values ($1,$2,$3,$4,$5,$6,$7,$8,$9,null): ERROR: new row for relation "records" violates check constraint "c_lowercase_name"
Mar 16 04:58:37 ns3 pdns_server: DETAIL: Failing row contains (6802, 5, 0A.client.nnn.fi, A, 192.168.0.10, 300, 0, null, f, null, t).
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Aborting possible open transaction for domain 'nnn.fi.' AXFR
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 Got a signal 11, attempting to print trace:
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0x18195b) [0x7fde2958495b]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(+0x35670) [0x7fde27948670]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/pgsql-9.5/lib/libpq.so.5(PQclear+0x69) [0x7fde272c0f49]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN15SPgSQLStatementD0Ev+0x5d) [0x7fde274f4f5d]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN11GSQLBackend14freeStatementsEv+0x199) [0x7fde294967d9]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/lib64/pdns/libgpgsqlbackend.so(_ZN13gPgSQLBackendD0Ev+0x34) [0x7fde274f1d24]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackend7cleanupEv+0x184) [0x7fde295d9f84]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN12UeberBackendD2Ev+0x2b) [0x7fde295da03b]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass4suckERK7DNSNameRKSs+0x5fc) [0x7fde295b211c]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(_ZN17CommunicatorClass19retrievalLoopThreadEv+0x136) [0x7fde294aded6]
Mar 16 04:58:37 ns3 pdns_server: Mar 16 10:58:37 /usr/sbin/pdns_server(+0xaea69) [0x7fde294b1a69]
Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libpthread.so.0(+0x7dc5) [0x7fde27cdbdc5]
Mar 16 04:58:38 ns3 pdns_server: Mar 16 10:58:37 /lib64/libc.so.6(clone+0x6d) [0x7fde27a0928d]
``` | defect | pdns crashes when database insert fails to postgres database lowercase constraint in the current pdns setup when the axfr transfer contains names which has uppercase letters the following happens mar pdns server mar transaction started for nnn fi mar pdns server mar unable to feed record during incoming axfr of nnn fi gsqlbackend unable to feed record fatal error during query insert into records content ttl prio type domain id disabled name ordername auth change date values null error new row for relation records violates check constraint c lowercase name mar pdns server detail failing row contains client nnn fi a null f null t mar pdns server mar aborting possible open transaction for domain nnn fi axfr mar pdns server mar got a signal attempting to print trace mar pdns server mar usr sbin pdns server mar pdns server mar libc so mar pdns server mar usr pgsql lib libpq so pqclear mar pdns server mar usr pdns libgpgsqlbackend so mar pdns server mar usr sbin pdns server mar pdns server mar usr pdns libgpgsqlbackend so mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar usr sbin pdns server mar pdns server mar libpthread so mar pdns server mar libc so clone | 1 |
7,356 | 2,610,364,949 | IssuesEvent | 2015-02-26 19:57:56 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | Config save failing | auto-migrated Priority-Medium tumblr Type-Defect | ```
What's the problem?
Wordpress.com and tumblr.com config fail to save. Login info has been confirmed.
What browser are you using?
Chrome + Windows
What version of ScribeFire are you running?
ScribeFire 4.1
```
-----
Original issue reported on code.google.com by `alexande...@sequoiavision.com` on 21 Dec 2012 at 3:12 | 1.0 | Config save failing - ```
What's the problem?
Wordpress.com and tumblr.com config fail to save. Login info has been confirmed.
What browser are you using?
Chrome + Windows
What version of ScribeFire are you running?
ScribeFire 4.1
```
-----
Original issue reported on code.google.com by `alexande...@sequoiavision.com` on 21 Dec 2012 at 3:12 | defect | config save failing what s the problem wordpress com and tumblr com config fail to save login info has been confirmed what browser are you using chrome windows what version of scribefire are you running scribefire original issue reported on code google com by alexande sequoiavision com on dec at | 1 |
30,666 | 6,219,850,217 | IssuesEvent | 2017-07-09 17:23:12 | pymc-devs/pymc3 | https://api.github.com/repos/pymc-devs/pymc3 | closed | sample_smc: AttributeError: 'NoneType' object has no attribute 'n_steps' | defects | ```
with pm.Model():
pm.NormalMixture('m',
mu=np.array([0., 1.]),
w=np.array([.6, .4]),
sd=np.array([.2, .2]))
step = pm.smc.SMC(n_chains=500)
trace = pm.smc.sample_smc(n_steps=1000, step=step, homepath='/tmp')
```
`traceplot(trace)` gives:

`sns.distplot(trace.get_values('m'));` gives:

| 1.0 | sample_smc: AttributeError: 'NoneType' object has no attribute 'n_steps' - ```
with pm.Model():
pm.NormalMixture('m',
mu=np.array([0., 1.]),
w=np.array([.6, .4]),
sd=np.array([.2, .2]))
step = pm.smc.SMC(n_chains=500)
trace = pm.smc.sample_smc(n_steps=1000, step=step, homepath='/tmp')
```
`traceplot(trace)` gives:

`sns.distplot(trace.get_values('m'));` gives:

| defect | sample smc attributeerror nonetype object has no attribute n steps with pm model pm normalmixture m mu np array w np array sd np array step pm smc smc n chains trace pm smc sample smc n steps step step homepath tmp traceplot trace gives sns distplot trace get values m gives | 1 |
81,309 | 30,791,832,407 | IssuesEvent | 2023-07-31 16:44:09 | vector-im/element-desktop | https://api.github.com/repos/vector-im/element-desktop | opened | MacOS: Unable to access microphone when try to call | T-Defect | ### Steps to reproduce
1. Upgrade the formerly working client to current version (up from 1.11.13) on a MacOS Ventura machine.
2. As soon as you start a call you receive the failure message. If you get a call there's not even the button to take it.
3. Downgrade to the last working version 1.11.13 restores the ability to call and take calls.
### Outcome
#### What did you expect?
I expected the Calls feature to keep working when upgrading to a more current version of Element.
#### What happened instead?
When I try to start a voice/video call, I get the following error and can not start the call:
```` Unable to access microphone
Call failed because microphone could not be accessed. Check that a microphone is plugged in and set up correctly.
````
### Operating system
MacOS
### Application version
Element higher than 1.11.13
### How did you install the app?
https://element.io
### Homeserver
docker image matrixdotorg/synapse:latest
### Will you send logs?
Yes | 1.0 | MacOS: Unable to access microphone when try to call - ### Steps to reproduce
1. Upgrade the formerly working client to current version (up from 1.11.13) on a MacOS Ventura machine.
2. As soon as you start a call you receive the failure message. If you get a call there's not even the button to take it.
3. Downgrade to the last working version 1.11.13 restores the ability to call and take calls.
### Outcome
#### What did you expect?
I expected the Calls feature to keep working when upgrading to a more current version of Element.
#### What happened instead?
When I try to start a voice/video call, I get the following error and can not start the call:
```` Unable to access microphone
Call failed because microphone could not be accessed. Check that a microphone is plugged in and set up correctly.
````
### Operating system
MacOS
### Application version
Element higher than 1.11.13
### How did you install the app?
https://element.io
### Homeserver
docker image matrixdotorg/synapse:latest
### Will you send logs?
Yes | defect | macos unable to access microphone when try to call steps to reproduce upgrade the formerly working client to current version up from on a macos ventura machine as soon as you start a call you receive the failure message if you get a call there s not even the button to take it downgrade to the last working version restores the ability to call and take calls outcome what did you expect i expected the calls feature to keep working when upgrading to a more current version of element what happened instead when i try to start a voice video call i get the following error and can not start the call unable to access microphone call failed because microphone could not be accessed check that a microphone is plugged in and set up correctly operating system macos application version element higher than how did you install the app homeserver docker image matrixdotorg synapse latest will you send logs yes | 1 |
433,111 | 30,313,058,391 | IssuesEvent | 2023-07-10 13:58:32 | SciTools/iris | https://api.github.com/repos/SciTools/iris | opened | Change the default version of the documentation to stable rather than latest on ReadTheDocs | Type: Documentation | ## 📚 Documentation
<!-- See https://scitools-iris.readthedocs.io/en/latest/ -->
<!-- Describe the issue or provide a suggestion for improving the Iris documentation -->
Currently the default version for the documentation on readthedocs is latest, which is the documentation for the next, unreleased version of Iris. This should be changed to the stable, which is the most recent release.
Because the unreleased documentation is the default, it is what is directed to when you visit the top of the domain ([scitools-iris.readthedocs.io](https://scitools-iris.readthedocs.io/)), as well as affecting the [canonical URL](https://docs.readthedocs.io/en/latest/canonical-urls.html#how-read-the-docs-generates-canonical-urls), which makes search engines point to the unreleased documentation in over the stable docs.
The current behaviour is both confusing, and [not recommended](https://docs.readthedocs.io/en/latest/guides/best-practice/links.html#questionable-practice). Fortunately it can be changed fairly easily in the ReadTheDocs web interface. The [setting](https://docs.readthedocs.io/en/latest/config-file/v2.html#migrating-from-the-web-interface) is `Admin` > `Advanced settings` > `Default version`. | 1.0 | Change the default version of the documentation to stable rather than latest on ReadTheDocs - ## 📚 Documentation
<!-- See https://scitools-iris.readthedocs.io/en/latest/ -->
<!-- Describe the issue or provide a suggestion for improving the Iris documentation -->
Currently the default version for the documentation on readthedocs is latest, which is the documentation for the next, unreleased version of Iris. This should be changed to the stable, which is the most recent release.
Because the unreleased documentation is the default, it is what is directed to when you visit the top of the domain ([scitools-iris.readthedocs.io](https://scitools-iris.readthedocs.io/)), as well as affecting the [canonical URL](https://docs.readthedocs.io/en/latest/canonical-urls.html#how-read-the-docs-generates-canonical-urls), which makes search engines point to the unreleased documentation in over the stable docs.
The current behaviour is both confusing, and [not recommended](https://docs.readthedocs.io/en/latest/guides/best-practice/links.html#questionable-practice). Fortunately it can be changed fairly easily in the ReadTheDocs web interface. The [setting](https://docs.readthedocs.io/en/latest/config-file/v2.html#migrating-from-the-web-interface) is `Admin` > `Advanced settings` > `Default version`. | non_defect | change the default version of the documentation to stable rather than latest on readthedocs 📚 documentation currently the default version for the documentation on readthedocs is latest which is the documentation for the next unreleased version of iris this should be changed to the stable which is the most recent release because the unreleased documentation is the default it is what is directed to when you visit the top of the domain as well as affecting the which makes search engines point to the unreleased documentation in over the stable docs the current behaviour is both confusing and fortunately it can be changed fairly easily in the readthedocs web interface the is admin advanced settings default version | 0 |
46,237 | 9,915,049,486 | IssuesEvent | 2019-06-28 15:48:36 | Badgerati/Pode | https://api.github.com/repos/Badgerati/Pode | opened | Split the flash function into multiple PowerShell functions | internal-code :hammer: | The `flash` function's actions should be split out into multiple PowerShell functions. Each of these functions should be exported in `pode.psd1`.
| Action | New Function | Returns |
| ------ | ------------ | ------- |
| `Add` | `Add-PodeFlashMessage` | `void` |
| `Clear` | `Clear-PodeFlashMessages` | `void` |
| `Get` | `Get-PodeFlashMessage` | `array` |
| `Keys` | `Get-PodeFlashMessageKeys` | `array` |
| `Remove` | `Remove-PodeFlashMessage` | `bool` |
Possible new functions:
* `Test-PodeFlashMessageKey` (test if a key is present) | 1.0 | Split the flash function into multiple PowerShell functions - The `flash` function's actions should be split out into multiple PowerShell functions. Each of these functions should be exported in `pode.psd1`.
| Action | New Function | Returns |
| ------ | ------------ | ------- |
| `Add` | `Add-PodeFlashMessage` | `void` |
| `Clear` | `Clear-PodeFlashMessages` | `void` |
| `Get` | `Get-PodeFlashMessage` | `array` |
| `Keys` | `Get-PodeFlashMessageKeys` | `array` |
| `Remove` | `Remove-PodeFlashMessage` | `bool` |
Possible new functions:
* `Test-PodeFlashMessageKey` (test if a key is present) | non_defect | split the flash function into multiple powershell functions the flash function s actions should be split out into multiple powershell functions each of these functions should be exported in pode action new function returns add add podeflashmessage void clear clear podeflashmessages void get get podeflashmessage array keys get podeflashmessagekeys array remove remove podeflashmessage bool possible new functions test podeflashmessagekey test if a key is present | 0 |
139,582 | 12,875,744,979 | IssuesEvent | 2020-07-11 00:31:09 | dotnet/diagnostics | https://api.github.com/repos/dotnet/diagnostics | opened | Add aks.ms link to diagnostics FAQ in SOS error messages | documentation dotnet-sos | And document in the FAQ the reasons for this message:
```
Failed to find runtime module (libcoreclr.so), 0x80004005
Extension commands need it in order to have something to do.
ClrStack failed
```
Some of the reasons for this message are:
1) the process or core dump hasn't loaded the .NET Core runtime yet.
2) lldb on a coredump wasn't started with the host program . `target modules list` doesn't display libcoreclr.so.
3) lldb on a coredump needs the correct version libcoreclr.so on disk. `target modules list` displays libcoreclr.so but SOS commands display this message. Use `dotnet-symbol --modules <coredump>` to download libcoreclr.so from the coredump.
| 1.0 | Add aks.ms link to diagnostics FAQ in SOS error messages - And document in the FAQ the reasons for this message:
```
Failed to find runtime module (libcoreclr.so), 0x80004005
Extension commands need it in order to have something to do.
ClrStack failed
```
Some of the reasons for this message are:
1) the process or core dump hasn't loaded the .NET Core runtime yet.
2) lldb on a coredump wasn't started with the host program . `target modules list` doesn't display libcoreclr.so.
3) lldb on a coredump needs the correct version libcoreclr.so on disk. `target modules list` displays libcoreclr.so but SOS commands display this message. Use `dotnet-symbol --modules <coredump>` to download libcoreclr.so from the coredump.
| non_defect | add aks ms link to diagnostics faq in sos error messages and document in the faq the reasons for this message failed to find runtime module libcoreclr so extension commands need it in order to have something to do clrstack failed some of the reasons for this message are the process or core dump hasn t loaded the net core runtime yet lldb on a coredump wasn t started with the host program target modules list doesn t display libcoreclr so lldb on a coredump needs the correct version libcoreclr so on disk target modules list displays libcoreclr so but sos commands display this message use dotnet symbol modules to download libcoreclr so from the coredump | 0 |
24,168 | 3,922,117,522 | IssuesEvent | 2016-04-22 03:49:32 | furushchev/alchemy-2 | https://api.github.com/repos/furushchev/alchemy-2 | closed | Weight 10 not working | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
bin/infer -i test.mln -e empty.db -r results.txt -q good
What is the expected output? What do you see instead?
expected (actually produced for weights 9.999999 and 10.000001)
good(A) 4.9995e-05
instead
good(A) 0.99995
What version of the product are you using? On what operating system?
Latest release (Jan 2013) on debian wheezy
Please provide any additional information below.
All other weights work just fine, but for some reason 10 doesn't work.
```
Original issue reported on code.google.com by `ads...@gmail.com` on 19 Jan 2014 at 9:11
Attachments:
* [test.mln](https://storage.googleapis.com/google-code-attachments/alchemy-2/issue-6/comment-0/test.mln)
| 1.0 | Weight 10 not working - ```
What steps will reproduce the problem?
bin/infer -i test.mln -e empty.db -r results.txt -q good
What is the expected output? What do you see instead?
expected (actually produced for weights 9.999999 and 10.000001)
good(A) 4.9995e-05
instead
good(A) 0.99995
What version of the product are you using? On what operating system?
Latest release (Jan 2013) on debian wheezy
Please provide any additional information below.
All other weights work just fine, but for some reason 10 doesn't work.
```
Original issue reported on code.google.com by `ads...@gmail.com` on 19 Jan 2014 at 9:11
Attachments:
* [test.mln](https://storage.googleapis.com/google-code-attachments/alchemy-2/issue-6/comment-0/test.mln)
| defect | weight not working what steps will reproduce the problem bin infer i test mln e empty db r results txt q good what is the expected output what do you see instead expected actually produced for weights and good a instead good a what version of the product are you using on what operating system latest release jan on debian wheezy please provide any additional information below all other weights work just fine but for some reason doesn t work original issue reported on code google com by ads gmail com on jan at attachments | 1 |
68,448 | 21,664,594,704 | IssuesEvent | 2022-05-07 01:53:13 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | New reactions picker sometimes goes off the top of the screen | T-Defect A-Reactions | I'll do a screenshot when I see it again - from memory I was reacting to a messages just about halfway up the visible timeline.
Firefox osx | 1.0 | New reactions picker sometimes goes off the top of the screen - I'll do a screenshot when I see it again - from memory I was reacting to a messages just about halfway up the visible timeline.
Firefox osx | defect | new reactions picker sometimes goes off the top of the screen i ll do a screenshot when i see it again from memory i was reacting to a messages just about halfway up the visible timeline firefox osx | 1 |
385,369 | 11,419,709,401 | IssuesEvent | 2020-02-03 08:37:23 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Exception handling issue in CacheCleanupTask | Priority/Highest Severity/Critical Type/Bug | We have the following code block in [1] to clear up caches that are expired or overflowed. There are multiple caches in the "caches" list and they are cleaned by invoking the `runCacheExpiry()` of each of them going through a loop.
```
try {
...
for (CacheImpl cache : caches) {
cache.runCacheExpiry();
}
}
} catch (IllegalStateException e) {
log.debug("Error occurred while running CacheCleanupTask", e);
```
If `cache.runCacheExpiry()` throws an exception, the whole cleanup task will break and skips clearing the rest of the caches after this exception. Therefore, it possible that some caches not cleaned up for a longer time and get overflowed. Cache eviction also happening in the same method `cache.runCacheExpiry()`.
Therefore, those non-cleaned caches do not accept new cache entries and different flows start to fail.
***Suggestion***
Catch IllegalStateException inside the loop itself and avoid failing the whole CacheCleanupTask
[1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/javax.cache/src/main/java/org/wso2/carbon/caching/impl/CacheCleanupTask.java#L47 | 1.0 | Exception handling issue in CacheCleanupTask - We have the following code block in [1] to clear up caches that are expired or overflowed. There are multiple caches in the "caches" list and they are cleaned by invoking the `runCacheExpiry()` of each of them going through a loop.
```
try {
...
for (CacheImpl cache : caches) {
cache.runCacheExpiry();
}
}
} catch (IllegalStateException e) {
log.debug("Error occurred while running CacheCleanupTask", e);
```
If `cache.runCacheExpiry()` throws an exception, the whole cleanup task will break and skips clearing the rest of the caches after this exception. Therefore, it possible that some caches not cleaned up for a longer time and get overflowed. Cache eviction also happening in the same method `cache.runCacheExpiry()`.
Therefore, those non-cleaned caches do not accept new cache entries and different flows start to fail.
***Suggestion***
Catch IllegalStateException inside the loop itself and avoid failing the whole CacheCleanupTask
[1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/javax.cache/src/main/java/org/wso2/carbon/caching/impl/CacheCleanupTask.java#L47 | non_defect | exception handling issue in cachecleanuptask we have the following code block in to clear up caches that are expired or overflowed there are multiple caches in the caches list and they are cleaned by invoking the runcacheexpiry of each of them going through a loop try for cacheimpl cache caches cache runcacheexpiry catch illegalstateexception e log debug error occurred while running cachecleanuptask e if cache runcacheexpiry throws an exception the whole cleanup task will break and skips clearing the rest of the caches after this exception therefore it possible that some caches not cleaned up for a longer time and get overflowed cache eviction also happening in the same method cache runcacheexpiry therefore those non cleaned caches do not accept new cache entries and different flows start to fail suggestion catch illegalstateexception inside the loop itself and avoid failing the whole cachecleanuptask | 0 |
234,664 | 7,724,418,864 | IssuesEvent | 2018-05-24 15:02:13 | test4gloirin/m | https://api.github.com/repos/test4gloirin/m | closed | 0000074:
replace getXXX/countXXX functions | Addressbook Feature Request low priority | **Reported by pschuele on 14 Jul 2008 15:17**
All getXXX/countXXX in json/controller/backend classes should be replaced by one get and one count function (see crm and task applications).
javascript frontend has to be updated.
| 1.0 | 0000074:
replace getXXX/countXXX functions - **Reported by pschuele on 14 Jul 2008 15:17**
All getXXX/countXXX in json/controller/backend classes should be replaced by one get and one count function (see crm and task applications).
javascript frontend has to be updated.
| non_defect | replace getxxx countxxx functions reported by pschuele on jul all getxxx countxxx in json controller backend classes should be replaced by one get and one count function see crm and task applications javascript frontend has to be updated | 0 |
33,009 | 6,994,840,804 | IssuesEvent | 2017-12-15 16:47:54 | contao/installation-bundle | https://api.github.com/repos/contao/installation-bundle | closed | The MySQL version check cannot determine the current version | defect | It appears that the [check for the MySQL version](https://github.com/contao/installation-bundle/commit/0514c650bce809bc8081badcdb1cda291f31eecf) may not always correctly determine the current version:

Tested with Contao 4.5.0-RC1 and `mysql Ver 14.14 Distrib 5.1.73, for redhat-linux-gnu (x86_64)` | 1.0 | The MySQL version check cannot determine the current version - It appears that the [check for the MySQL version](https://github.com/contao/installation-bundle/commit/0514c650bce809bc8081badcdb1cda291f31eecf) may not always correctly determine the current version:

Tested with Contao 4.5.0-RC1 and `mysql Ver 14.14 Distrib 5.1.73, for redhat-linux-gnu (x86_64)` | defect | the mysql version check cannot determine the current version it appears that the may not always correctly determine the current version tested with contao and mysql ver distrib for redhat linux gnu | 1 |
289,255 | 24,971,143,150 | IssuesEvent | 2022-11-02 01:21:27 | CMU-313/fall-22-hw4-team-sweg | https://api.github.com/repos/CMU-313/fall-22-hw4-team-sweg | closed | Test API to delete a model | test | Currently, there are no tests that make sure the delete function works as expected and takes in an appropriate model_id. I will add tests to test_models.py that will assert that the delete works as expected. | 1.0 | Test API to delete a model - Currently, there are no tests that make sure the delete function works as expected and takes in an appropriate model_id. I will add tests to test_models.py that will assert that the delete works as expected. | non_defect | test api to delete a model currently there are no tests that make sure the delete function works as expected and takes in an appropriate model id i will add tests to test models py that will assert that the delete works as expected | 0 |
30,734 | 6,261,182,665 | IssuesEvent | 2017-07-14 23:00:21 | blong42/clearsilver | https://api.github.com/repos/blong42/clearsilver | closed | problem importing neo_cgi : missing crc32 | auto-migrated Priority-Medium Type-Defect | ```
$ cd /usr/local/lib/python2.4/site-packages/site-packages$ python -c "import
neo_cgi"
Traceback (most recent call last):
File "<string>", line 1, in ?
ImportError: ./neo_cgi.so: undefined symbol: crc32
I am sure you assumed it would come from zlib.h ... not on my system!
I imagine this has to do with Config.CompressionEnabled.
Linux 3.2.0-29-generic #46-Ubuntu SMP Fri Jul 27 17:03:23 UTC 2012 x86_64
x86_64 x86_64 GNU/Linux
```
Original issue reported on code.google.com by `brianh...@gmail.com` on 16 Aug 2012 at 8:47
| 1.0 | problem importing neo_cgi : missing crc32 - ```
$ cd /usr/local/lib/python2.4/site-packages/site-packages$ python -c "import
neo_cgi"
Traceback (most recent call last):
File "<string>", line 1, in ?
ImportError: ./neo_cgi.so: undefined symbol: crc32
I am sure you assumed it would come from zlib.h ... not on my system!
I imagine this has to do with Config.CompressionEnabled.
Linux 3.2.0-29-generic #46-Ubuntu SMP Fri Jul 27 17:03:23 UTC 2012 x86_64
x86_64 x86_64 GNU/Linux
```
Original issue reported on code.google.com by `brianh...@gmail.com` on 16 Aug 2012 at 8:47
| defect | problem importing neo cgi missing cd usr local lib site packages site packages python c import neo cgi traceback most recent call last file line in importerror neo cgi so undefined symbol i am sure you assumed it would come from zlib h not on my system i imagine this has to do with config compressionenabled linux generic ubuntu smp fri jul utc gnu linux original issue reported on code google com by brianh gmail com on aug at | 1 |
26,802 | 4,789,119,292 | IssuesEvent | 2016-10-30 22:17:26 | belangeo/pyo | https://api.github.com/repos/belangeo/pyo | closed | Jack often zombifies on Ubuntu | auto-migrated Priority-Medium Type-Defect | ```
Run any pyo script with Jack support.
We often see "JACK error: zombified - calling shutdown handler".
Sometimes it's OK, sometimes it's not.
It's on Ubuntu Lucid, Natty, etc.
```
Original issue reported on code.google.com by `que...@gmail.com` on 21 Nov 2011 at 7:02
| 1.0 | Jack often zombifies on Ubuntu - ```
Run any pyo script with Jack support.
We often see "JACK error: zombified - calling shutdown handler".
Sometimes it's OK, sometimes it's not.
It's on Ubuntu Lucid, Natty, etc.
```
Original issue reported on code.google.com by `que...@gmail.com` on 21 Nov 2011 at 7:02
| defect | jack often zombifies on ubuntu run any pyo script with jack support we often see jack error zombified calling shutdown handler sometimes it s ok sometimes it s not it s on ubuntu lucid natty etc original issue reported on code google com by que gmail com on nov at | 1 |
407,074 | 11,906,191,221 | IssuesEvent | 2020-03-30 19:55:24 | gsoft-inc/sg-orbit | https://api.github.com/repos/gsoft-inc/sg-orbit | closed | A select item with an icon doesn't follow text / icon ratio | bug components high priority theming | <!--
Please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo.
If the bug prevent your team from releasing please add the tag "blocking" to your issue.
-->
### Describe the bug

A select item icon size adapt to the select size but the text doesn't. Therefore, it breaks our standard text /icon ratio.
Available options:
- Don't adapt the icon size to the select size
- Adapt the text size to the select size | 1.0 | A select item with an icon doesn't follow text / icon ratio - <!--
Please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo.
If the bug prevent your team from releasing please add the tag "blocking" to your issue.
-->
### Describe the bug

A select item icon size adapt to the select size but the text doesn't. Therefore, it breaks our standard text /icon ratio.
Available options:
- Don't adapt the icon size to the select size
- Adapt the text size to the select size | non_defect | a select item with an icon doesn t follow text icon ratio please note that your issue will be fixed much faster if you includes the exact reproduction steps and a demo if the bug prevent your team from releasing please add the tag blocking to your issue describe the bug a select item icon size adapt to the select size but the text doesn t therefore it breaks our standard text icon ratio available options don t adapt the icon size to the select size adapt the text size to the select size | 0 |
48,907 | 13,184,772,576 | IssuesEvent | 2020-08-12 20:03:58 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/423
, reported by claudio.kopper and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:36",
"description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"_ts": "1351704816000000",
"component": "combo reconstruction",
"summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional",
"priority": "normal",
"keywords": "",
"time": "2012-06-25T00:59:43",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/423
, reported by claudio.kopper and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:36",
"description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"_ts": "1351704816000000",
"component": "combo reconstruction",
"summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional",
"priority": "normal",
"keywords": "",
"time": "2012-06-25T00:59:43",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | sim services propagatorserviceutils propagate broken non functional trac migrated from reported by claudio kopper and owned by olivas json status closed changetime description i d like to commit the following attached patch with the following commit message n npropagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy n reporter claudio kopper cc resolution fixed ts component combo reconstruction summary sim services propagatorserviceutils propagate broken non functional priority normal keywords time milestone owner olivas type defect | 1 |
375,963 | 11,136,509,261 | IssuesEvent | 2019-12-20 16:46:03 | SynBioDex/SBOLExplorer | https://api.github.com/repos/SynBioDex/SBOLExplorer | closed | Create a separate SBOLExplorer log file | enhancement priority | This should be viewable on the SBOLExplorer admin page | 1.0 | Create a separate SBOLExplorer log file - This should be viewable on the SBOLExplorer admin page | non_defect | create a separate sbolexplorer log file this should be viewable on the sbolexplorer admin page | 0 |
27,284 | 4,957,388,172 | IssuesEvent | 2016-12-02 04:09:36 | google/gitiles | https://api.github.com/repos/google/gitiles | reopened | Add option for embedded (iframe) Gitiles views. | Priority-Medium Status-New Type-Defect | Originally reported on Google Code with ID 65
```
It is useful for applications wishing to embed Gitiles data to have access to a trimmed-down
form of Gitiles pages that omit specifically the Gitiles header and footer, focusing
only on the requested content. This is useful for applications wishing to embed Gitiles
data in (for example) an iframe.
The ability to render a embedded view should be independent of the content of that
view. In other words, the parameter should not use the "pretty" parameter. Instead,
a "view" parameter is proposed. The default "view" value is the current view. However,
alternative views can be supplied. In this case, the iframe-embeddable view would use
"?view=embed".
```
Reported by `None` on 2014-09-17 16:23:32
| 1.0 | Add option for embedded (iframe) Gitiles views. - Originally reported on Google Code with ID 65
```
It is useful for applications wishing to embed Gitiles data to have access to a trimmed-down
form of Gitiles pages that omit specifically the Gitiles header and footer, focusing
only on the requested content. This is useful for applications wishing to embed Gitiles
data in (for example) an iframe.
The ability to render a embedded view should be independent of the content of that
view. In other words, the parameter should not use the "pretty" parameter. Instead,
a "view" parameter is proposed. The default "view" value is the current view. However,
alternative views can be supplied. In this case, the iframe-embeddable view would use
"?view=embed".
```
Reported by `None` on 2014-09-17 16:23:32
| defect | add option for embedded iframe gitiles views originally reported on google code with id it is useful for applications wishing to embed gitiles data to have access to a trimmed down form of gitiles pages that omit specifically the gitiles header and footer focusing only on the requested content this is useful for applications wishing to embed gitiles data in for example an iframe the ability to render a embedded view should be independent of the content of that view in other words the parameter should not use the pretty parameter instead a view parameter is proposed the default view value is the current view however alternative views can be supplied in this case the iframe embeddable view would use view embed reported by none on | 1 |
79 | 2,504,337,062 | IssuesEvent | 2015-01-10 02:59:32 | eleybourn/Book-Catalogue | https://api.github.com/repos/eleybourn/Book-Catalogue | closed | Fit/finish - wrong heading for add book by title/author (search internet) | Defect In Beta | Adding a book by title and author (search internet) has the heading of ISBN search. | 1.0 | Fit/finish - wrong heading for add book by title/author (search internet) - Adding a book by title and author (search internet) has the heading of ISBN search. | defect | fit finish wrong heading for add book by title author search internet adding a book by title and author search internet has the heading of isbn search | 1 |
64,104 | 18,213,973,278 | IssuesEvent | 2021-09-30 00:14:25 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | ExtraNodesetGenerator can’t find correct node using coord when mesh dimension lower than spatial dimension | T: defect P: normal | ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0.
https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| 1.0 | ExtraNodesetGenerator can’t find correct node using coord when mesh dimension lower than spatial dimension - ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0.
https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| defect | extranodesetgenerator can’t find correct node using coord when mesh dimension lower than spatial dimension bug description when the mesh dimension is lower than spatial dimension say mesh dimension is spatial dimension is the code will only copy coordinate components of “c” into “p” which is not correct since c may not be steps to reproduce impact | 1 |
24,370 | 3,968,652,748 | IssuesEvent | 2016-05-03 20:26:10 | lester88a/snova | https://api.github.com/repos/lester88a/snova | closed | 新版0.23.0 Linux(32位)编译后运行问题? | auto-migrated Priority-Medium Type-Defect | ```
环境: v0.23.0 ubuntu14.04 x32
1-因为没有看到32位版本,自己编译第一次没通过,删掉图标�
��以了OpenShift所有功能正常,Heroku封了不是本应用的问题。但
有个疑问:运行在后台通过log可以查看运行情况,但退出怎��
�办?我是直接klii了。是否我哪里没有搞对还是没办法自己结
束进程?
2-其实老版本java部署在一些支持java的空间很不错,比较稳定�
��宜被封。另外老版本的user_spac.json,user_hosts.conf文件是否不�
��作用了?
3-snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov
a的原始作者吗,很崇拜!
```
Original issue reported on code.google.com by `bob.yang...@gmail.com` on 30 Oct 2014 at 2:48 | 1.0 | 新版0.23.0 Linux(32位)编译后运行问题? - ```
环境: v0.23.0 ubuntu14.04 x32
1-因为没有看到32位版本,自己编译第一次没通过,删掉图标�
��以了OpenShift所有功能正常,Heroku封了不是本应用的问题。但
有个疑问:运行在后台通过log可以查看运行情况,但退出怎��
�办?我是直接klii了。是否我哪里没有搞对还是没办法自己结
束进程?
2-其实老版本java部署在一些支持java的空间很不错,比较稳定�
��宜被封。另外老版本的user_spac.json,user_hosts.conf文件是否不�
��作用了?
3-snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov
a的原始作者吗,很崇拜!
```
Original issue reported on code.google.com by `bob.yang...@gmail.com` on 30 Oct 2014 at 2:48 | defect | linux( )编译后运行问题? 环境: ,自己编译第一次没通过,删掉图标� ��以了openshift所有功能正常,heroku封了不是本应用的问题。但 有个疑问:运行在后台通过log可以查看运行情况,但退出怎�� �办?我是直接klii了。是否我哪里没有搞对还是没办法自己结 束进程? 其实老版本java部署在一些支持java的空间很不错,比较稳定� ��宜被封。另外老版本的user spac json,user hosts conf文件是否不� ��作用了? snova的确非常优秀非今天一些应用可比,想知道yinqiwen是gsnov a的原始作者吗,很崇拜! original issue reported on code google com by bob yang gmail com on oct at | 1 |
73,258 | 24,527,706,886 | IssuesEvent | 2022-10-11 14:14:40 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | [Server] one_result_per_user_per_wu potentially broken | C: Server - Scheduler P: Minor R: worksforme T: Defect E: to be determined | **Describe the bug**
The server configuration option `one_result_per_user_per_wu` seems to not function as expected. I've seen 3 separate BOINC servers that have this option enabled and some users get two tasks from the same workunit. If the `one_result_per_host_per_wu` option is not enabled, a single host can get both tasks from a workunit:

I spoke with the SiDock@home admin and, at the time of that log entry, `one_result_per_user_per_wu` was enabled but `one_result_per_host_per_wu` was not.
**Affected Projects**
- Minecraft@Home
- SiDock@home
- my own private BOINC instance
**Steps To Reproduce**
1. Enable the rule `one_result_per_user_per_wu` in config.xml.
2. Wait for some new tasks to be distributed.
3. Go to the ops panel, click 'Workunits' under 'Browse database', then paste this into the 'additional clauses' box: `id in (select r.workunitid from result r join result r2 on r.workunitid = r2.workunitid and r.hostid <> r2.hostid and r.userid = r2.userid)`
4. Alternatively, in a clean BOINC server test environment, simply attach one host and watch as it gets both replications.
**Expected behavior**
The query should return no results - hosts should not get two results from the same workunit.
**Screenshots**
my BOINC instance:

Minecraft@Home:

SiDock@home:

**System Information**
- OS: Debian 10 Buster (used by my private BOINC server)
- BOINC Version: 1.2.0 (https://github.com/BOINC/boinc/tree/server_release/1/1.2) (version used by both my server and Minecraft@Home)
- BOINC Version: 1.1.0 (version used by SiDock@home)
**Additional Context**
Minecraft@Home config.xml: https://github.com/minecrafthome/minecrafthome/blob/master/images/makeproject/project/config.xml#L35
Feel free to ask for additional clarification if I missed anything!
| 1.0 | [Server] one_result_per_user_per_wu potentially broken - **Describe the bug**
The server configuration option `one_result_per_user_per_wu` seems to not function as expected. I've seen 3 separate BOINC servers that have this option enabled and some users get two tasks from the same workunit. If the `one_result_per_host_per_wu` option is not enabled, a single host can get both tasks from a workunit:

I spoke with the SiDock@home admin and, at the time of that log entry, `one_result_per_user_per_wu` was enabled but `one_result_per_host_per_wu` was not.
**Affected Projects**
- Minecraft@Home
- SiDock@home
- my own private BOINC instance
**Steps To Reproduce**
1. Enable the rule `one_result_per_user_per_wu` in config.xml.
2. Wait for some new tasks to be distributed.
3. Go to the ops panel, click 'Workunits' under 'Browse database', then paste this into the 'additional clauses' box: `id in (select r.workunitid from result r join result r2 on r.workunitid = r2.workunitid and r.hostid <> r2.hostid and r.userid = r2.userid)`
4. Alternatively, in a clean BOINC server test environment, simply attach one host and watch as it gets both replications.
**Expected behavior**
The query should return no results - hosts should not get two results from the same workunit.
**Screenshots**
my BOINC instance:

Minecraft@Home:

SiDock@home:

**System Information**
- OS: Debian 10 Buster (used by my private BOINC server)
- BOINC Version: 1.2.0 (https://github.com/BOINC/boinc/tree/server_release/1/1.2) (version used by both my server and Minecraft@Home)
- BOINC Version: 1.1.0 (version used by SiDock@home)
**Additional Context**
Minecraft@Home config.xml: https://github.com/minecrafthome/minecrafthome/blob/master/images/makeproject/project/config.xml#L35
Feel free to ask for additional clarification if I missed anything!
| defect | one result per user per wu potentially broken describe the bug the server configuration option one result per user per wu seems to not function as expected i ve seen separate boinc servers that have this option enabled and some users get two tasks from the same workunit if the one result per host per wu option is not enabled a single host can get both tasks from a workunit i spoke with the sidock home admin and at the time of that log entry one result per user per wu was enabled but one result per host per wu was not affected projects minecraft home sidock home my own private boinc instance steps to reproduce enable the rule one result per user per wu in config xml wait for some new tasks to be distributed go to the ops panel click workunits under browse database then paste this into the additional clauses box id in select r workunitid from result r join result on r workunitid workunitid and r hostid hostid and r userid userid alternatively in a clean boinc server test environment simply attach one host and watch as it gets both replications expected behavior the query should return no results hosts should not get two results from the same workunit screenshots my boinc instance minecraft home sidock home system information os debian buster used by my private boinc server boinc version version used by both my server and minecraft home boinc version version used by sidock home additional context minecraft home config xml feel free to ask for additional clarification if i missed anything | 1 |
174,173 | 27,589,179,489 | IssuesEvent | 2023-03-08 22:30:06 | NCIOCPL/cgov-digital-platform | https://api.github.com/repos/NCIOCPL/cgov-digital-platform | closed | update ncids version to use uswds normalize | Drupal - Redesign | ## Issue description
Bump package version of ncids to include uswds normalize file
https://github.com/NCIOCPL/ncids/issues/738 | 1.0 | update ncids version to use uswds normalize - ## Issue description
Bump package version of ncids to include uswds normalize file
https://github.com/NCIOCPL/ncids/issues/738 | non_defect | update ncids version to use uswds normalize issue description bump package version of ncids to include uswds normalize file | 0 |
365,279 | 10,780,354,687 | IssuesEvent | 2019-11-04 12:48:37 | emsec/hal | https://api.github.com/repos/emsec/hal | closed | Feature: Allow manual selection of module colors | Priority: Medium Status: Abandoned Type: Feature Request | HAL sometimes selects two colors very close to each other when creating many modules, making the different modules hard to see. I suggest a manual override for that color by:
- passing the color in the python command for creating a module
- right-clicking on a module in the navigation view | 1.0 | Feature: Allow manual selection of module colors - HAL sometimes selects two colors very close to each other when creating many modules, making the different modules hard to see. I suggest a manual override for that color by:
- passing the color in the python command for creating a module
- right-clicking on a module in the navigation view | non_defect | feature allow manual selection of module colors hal sometimes selects two colors very close to each other when creating many modules making the different modules hard to see i suggest a manual override for that color by passing the color in the python command for creating a module right clicking on a module in the navigation view | 0 |
51,251 | 13,207,402,164 | IssuesEvent | 2020-08-14 22:58:07 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | environment should be checked for PATH, etc. clobbering at cmake time (Trac #97) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/97">https://code.icecube.wisc.edu/projects/icecube/ticket/97</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"_ts": "1194753078000000",
"description": "",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2007-08-23T23:27:36",
"component": "cmake",
"summary": "environment should be checked for PATH, etc. clobbering at cmake time",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | environment should be checked for PATH, etc. clobbering at cmake time (Trac #97) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/97">https://code.icecube.wisc.edu/projects/icecube/ticket/97</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"_ts": "1194753078000000",
"description": "",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2007-08-23T23:27:36",
"component": "cmake",
"summary": "environment should be checked for PATH, etc. clobbering at cmake time",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | environment should be checked for path etc clobbering at cmake time trac migrated from json status closed changetime ts description reporter troy cc resolution fixed time component cmake summary environment should be checked for path etc clobbering at cmake time priority normal keywords milestone owner troy type defect | 1 |
171,191 | 27,081,820,294 | IssuesEvent | 2023-02-14 14:30:58 | hypha-dao/hypha_wallet | https://api.github.com/repos/hypha-dao/hypha_wallet | opened | UI review | Design | Dark mode & Light mode
- [ ] Scan QR banner
- [ ] Memo in transaction history (and detail)
- [ ] BottomNav menu: can we have more box-shadow? (see here)
- [ ] Transaction signing - action title
- [ ] transaction expires in
- [ ] Handle “while” sliding.
- [ ] Bottom Sheets
- [ ] Former Bottom sheet “transaction fail”
- [ ] Import account
- [ ] toggle::Active use Main app colour gradient (not green), and use the same component all over the app (also fix size pls if possible)
| 1.0 | UI review - Dark mode & Light mode
- [ ] Scan QR banner
- [ ] Memo in transaction history (and detail)
- [ ] BottomNav menu: can we have more box-shadow? (see here)
- [ ] Transaction signing - action title
- [ ] transaction expires in
- [ ] Handle “while” sliding.
- [ ] Bottom Sheets
- [ ] Former Bottom sheet “transaction fail”
- [ ] Import account
- [ ] toggle::Active use Main app colour gradient (not green), and use the same component all over the app (also fix size pls if possible)
| non_defect | ui review dark mode light mode scan qr banner memo in transaction history and detail bottomnav menu can we have more box shadow see here transaction signing action title transaction expires in handle “while” sliding bottom sheets former bottom sheet “transaction fail” import account toggle active use main app colour gradient not green and use the same component all over the app also fix size pls if possible | 0 |
33,717 | 7,201,510,337 | IssuesEvent | 2018-02-05 22:54:02 | otros-systems/otroslogviewer | https://api.github.com/repos/otros-systems/otroslogviewer | closed | OutOfMemory on continues tailing on big files. | Priority-Medium Type-Defect auto-migrated | ```
What steps will reproduce the problem?
1. Use stock OLV 1.2, choose "tail log with type detection". Choose log4j log
that is continuesly growing.
2. Wait until it reaches heap limit.
3. When it reaches memory threshold, application crashes with
"java.lang.OutOfMemoryError: Java heap ...". Sometimes even error boxes are not
drawed. (so cannot include exact error currently), application is unusable from
this point. Nothing at screen gets repainted, nothing reacts to clicks.
What is the expected output? What do you see instead?
a) Expected output usual application behavior tailing logs. (preffered)
b) Grid with log lines cleared to free up memory. Notification is shown
somewhere (which would NOT require to click ok to proceed) about error cause.
This needed when log watching is left for overnight. It should not crash.
What version of the product are you using? On what operating system?
OLV 1.2,
java version "1.7.0_67"
Java(TM) SE Runtime Environment (build 1.7.0_67-b01)
Java HotSpot(TM) 64-Bit Server VM (build 24.65-b04, mixed mode)
Windows 7x64 (Microsoft Windows [Version 6.1.7601])
Please provide any additional information below.
Problem could be repeated on Windows Server 20xx OS.
Desired fix is to be able to select option in settings: default size of
"current log window".
That is number of lines that are actually created in grid and occupies memory.
That is if window is 1000 lines, then when it comes 1002 line, first two lines
(1,2) are deleted from "current window" and freed from memory. if i use scroll
bar to go to the begining of the file, then "current window" is lines 1-1000,
if i go to the last lines, than "current window" becomes 3-1002 lines.
```
Original issue reported on code.google.com by `jaun...@gmail.com` on 5 Nov 2014 at 7:30
| 1.0 | OutOfMemory on continues tailing on big files. - ```
What steps will reproduce the problem?
1. Use stock OLV 1.2, choose "tail log with type detection". Choose log4j log
that is continuesly growing.
2. Wait until it reaches heap limit.
3. When it reaches memory threshold, application crashes with
"java.lang.OutOfMemoryError: Java heap ...". Sometimes even error boxes are not
drawed. (so cannot include exact error currently), application is unusable from
this point. Nothing at screen gets repainted, nothing reacts to clicks.
What is the expected output? What do you see instead?
a) Expected output usual application behavior tailing logs. (preffered)
b) Grid with log lines cleared to free up memory. Notification is shown
somewhere (which would NOT require to click ok to proceed) about error cause.
This needed when log watching is left for overnight. It should not crash.
What version of the product are you using? On what operating system?
OLV 1.2,
java version "1.7.0_67"
Java(TM) SE Runtime Environment (build 1.7.0_67-b01)
Java HotSpot(TM) 64-Bit Server VM (build 24.65-b04, mixed mode)
Windows 7x64 (Microsoft Windows [Version 6.1.7601])
Please provide any additional information below.
Problem could be repeated on Windows Server 20xx OS.
Desired fix is to be able to select option in settings: default size of
"current log window".
That is number of lines that are actually created in grid and occupies memory.
That is if window is 1000 lines, then when it comes 1002 line, first two lines
(1,2) are deleted from "current window" and freed from memory. if i use scroll
bar to go to the begining of the file, then "current window" is lines 1-1000,
if i go to the last lines, than "current window" becomes 3-1002 lines.
```
Original issue reported on code.google.com by `jaun...@gmail.com` on 5 Nov 2014 at 7:30
| defect | outofmemory on continues tailing on big files what steps will reproduce the problem use stock olv choose tail log with type detection choose log that is continuesly growing wait until it reaches heap limit when it reaches memory threshold application crashes with java lang outofmemoryerror java heap sometimes even error boxes are not drawed so cannot include exact error currently application is unusable from this point nothing at screen gets repainted nothing reacts to clicks what is the expected output what do you see instead a expected output usual application behavior tailing logs preffered b grid with log lines cleared to free up memory notification is shown somewhere which would not require to click ok to proceed about error cause this needed when log watching is left for overnight it should not crash what version of the product are you using on what operating system olv java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode windows microsoft windows please provide any additional information below problem could be repeated on windows server os desired fix is to be able to select option in settings default size of current log window that is number of lines that are actually created in grid and occupies memory that is if window is lines then when it comes line first two lines are deleted from current window and freed from memory if i use scroll bar to go to the begining of the file then current window is lines if i go to the last lines than current window becomes lines original issue reported on code google com by jaun gmail com on nov at | 1 |
62,400 | 17,023,915,141 | IssuesEvent | 2021-07-03 04:32:08 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | missing boundaries for continents in nominatim search | Component: nominatim Priority: minor Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 1.41pm, Tuesday, 16th December 2014]**
when searching for australia, europe...
e.g. [http://nominatim.openstreetmap.org/search.php?q=Europe&polygon=1]
it finds them on the map but does not show boundaries. | 1.0 | missing boundaries for continents in nominatim search - **[Submitted to the original trac issue database at 1.41pm, Tuesday, 16th December 2014]**
when searching for australia, europe...
e.g. [http://nominatim.openstreetmap.org/search.php?q=Europe&polygon=1]
it finds them on the map but does not show boundaries. | defect | missing boundaries for continents in nominatim search when searching for australia europe e g it finds them on the map but does not show boundaries | 1 |
176,059 | 14,549,172,041 | IssuesEvent | 2020-12-16 03:07:09 | liamlau/individual-project | https://api.github.com/repos/liamlau/individual-project | opened | Write Documentation Requirements Wiki Page | documentation | At https://github.com/liamlau/individual-project/wiki/Documentation-%7C-Requirements-(MoSCoW)
Take the OneNote notes I have and write them here in a better format. | 1.0 | Write Documentation Requirements Wiki Page - At https://github.com/liamlau/individual-project/wiki/Documentation-%7C-Requirements-(MoSCoW)
Take the OneNote notes I have and write them here in a better format. | non_defect | write documentation requirements wiki page at take the onenote notes i have and write them here in a better format | 0 |
73,393 | 24,607,520,654 | IssuesEvent | 2022-10-14 17:43:19 | idaholab/HERON | https://api.github.com/repos/idaholab/HERON | closed | [DEFECT] Mac Documentation Failed | defect | --------
Defect Description
--------
**Describe the defect**
##### What did you expect to see happen?
The user manual to be built
##### What did you see instead?
Import errors stopping the build process
<img width="1016" alt="image" src="https://user-images.githubusercontent.com/74023858/195851747-b024752a-fdb7-43f4-ac34-7ded3aeb8e9c.png">
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1. Update and build raven heron and teal
2. Activate raven_libraries environment
3. Try to build HERON documentation
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| 1.0 | [DEFECT] Mac Documentation Failed - --------
Defect Description
--------
**Describe the defect**
##### What did you expect to see happen?
The user manual to be built
##### What did you see instead?
Import errors stopping the build process
<img width="1016" alt="image" src="https://user-images.githubusercontent.com/74023858/195851747-b024752a-fdb7-43f4-ac34-7ded3aeb8e9c.png">
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1. Update and build raven heron and teal
2. Activate raven_libraries environment
3. Try to build HERON documentation
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| defect | mac documentation failed defect description describe the defect what did you expect to see happen the user manual to be built what did you see instead import errors stopping the build process img width alt image src do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior update and build raven heron and teal activate raven libraries environment try to build heron documentation screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided | 1 |
2,437 | 3,683,179,959 | IssuesEvent | 2016-02-24 13:03:08 | elmsln/elmsln | https://api.github.com/repos/elmsln/elmsln | closed | support for letsencrypt | enhancement infrastructure security related | http://letsencrypt.org is awesome and would allow us to (hopefully) automate certificate granting | 1.0 | support for letsencrypt - http://letsencrypt.org is awesome and would allow us to (hopefully) automate certificate granting | non_defect | support for letsencrypt is awesome and would allow us to hopefully automate certificate granting | 0 |
174,937 | 21,300,580,151 | IssuesEvent | 2022-04-15 02:11:15 | turkdevops/vue-devtools | https://api.github.com/repos/turkdevops/vue-devtools | opened | CVE-2021-43138 (High) detected in async-2.6.1.tgz | security vulnerability | ## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-2.6.1.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.1.tgz">https://registry.npmjs.org/async/-/async-2.6.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- cypress-3.8.3.tgz (Root Library)
- getos-3.1.1.tgz
- :x: **async-2.6.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dependabot/npm_and_yarn/electron-9.4.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution (async): 3.2.2</p>
<p>Direct dependency fix Resolution (cypress): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-43138 (High) detected in async-2.6.1.tgz - ## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-2.6.1.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.1.tgz">https://registry.npmjs.org/async/-/async-2.6.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- cypress-3.8.3.tgz (Root Library)
- getos-3.1.1.tgz
- :x: **async-2.6.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dependabot/npm_and_yarn/electron-9.4.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution (async): 3.2.2</p>
<p>Direct dependency fix Resolution (cypress): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in async tgz cve high severity vulnerability vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file package json path to vulnerable library node modules async package json dependency hierarchy cypress tgz root library getos tgz x async tgz vulnerable library found in base branch dependabot npm and yarn electron vulnerability details a vulnerability exists in async through fixed in which could let a malicious user obtain privileges via the mapvalues method publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async direct dependency fix resolution cypress step up your open source security game with whitesource | 0 |
20,580 | 3,385,468,587 | IssuesEvent | 2015-11-27 11:41:57 | contao/core | https://api.github.com/repos/contao/core | closed | \Request and relative 301 location | defect | Requesting entries from a sitemap http://www.heise.de/sitemap.xml
Those entries are redirected to a relative URL, which is wrong handled by the Request::send function
Array
(
[scheme] => http
[host] => www.heise.de
[path] => /meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html
)
Redirect goes to :
Array
(
[path] => /newsticker/meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html
)
that means, the redirect here (https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Request.php#L443) has to check, if scheme, host, port exists and rebuild the URL | 1.0 | \Request and relative 301 location - Requesting entries from a sitemap http://www.heise.de/sitemap.xml
Those entries are redirected to a relative URL, which is wrong handled by the Request::send function
Array
(
[scheme] => http
[host] => www.heise.de
[path] => /meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html
)
Redirect goes to :
Array
(
[path] => /newsticker/meldung/Hickhack-um-Virtual-Reality-Apps-fuer-Samsungs-Gear-VR-2631443.html
)
that means, the redirect here (https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Request.php#L443) has to check, if scheme, host, port exists and rebuild the URL | defect | request and relative location requesting entries from a sitemap those entries are redirected to a relative url which is wrong handled by the request send function array http meldung hickhack um virtual reality apps fuer samsungs gear vr html redirect goes to array newsticker meldung hickhack um virtual reality apps fuer samsungs gear vr html that means the redirect here has to check if scheme host port exists and rebuild the url | 1 |
61,056 | 17,023,590,485 | IssuesEvent | 2021-07-03 02:48:47 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | OSM Server Side Script: sketch_line and sketch_route produce erratic output | Component: utils Priority: major Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 11.22pm, Friday, 14th May 2010]**
sketch_line (sketch_route is also affected) often produces erratic graphics. I have observed two issues:
1. apparently, if one of the stop members of the relation is an area tagged as amenity=bus_station, it is omitted completely. For example:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=80&style=padua]
consisting of relations:
[http://www.openstreetmap.org/browse/relation/611446]
[http://www.openstreetmap.org/browse/relation/611538]
The first stop (Bonola) is tagged as described above and missing from the sketch. This seems incorrect to me as, to my knowledge, bus_station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms.
2. Trolleybus lines (route=trolleybus) are not rendered at all - attempting to sketch them results in "No relation found".
To reproduce:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=92&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/913880]
[http://www.openstreetmap.org/browse/relation/913907]
3. in some cases incorrect stop names are displayed - names are mixed up or the same stop name is displayed multiple times. The graph itself seems correct (though I have not verified this in too much detail).
Mild example:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=19&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/661028]
[http://www.openstreetmap.org/browse/relation/661000]
Here just one stop name is wrong: the 11th (from Roserio) is displayed as "Duomo" when in fact it should be "Piazzale Accursio". (Duomo is the last stop.) Stops at which only one direction stops seem to be particular prone to this error (though it happens to all).
A more extreme one would be:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=78&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/420649]
[http://www.openstreetmap.org/browse/relation/445904]
This one is botched beyond recognition and usability: names are either missing completely, or the same name is repeated over and over multiple times.
And finally:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=69&style=padua]
two alternative routes, resulting in a total of four relations:
[http://www.openstreetmap.org/browse/relation/536964]
[http://www.openstreetmap.org/browse/relation/537040]
[http://www.openstreetmap.org/browse/relation/536992]
[http://www.openstreetmap.org/browse/relation/537008] | 1.0 | OSM Server Side Script: sketch_line and sketch_route produce erratic output - **[Submitted to the original trac issue database at 11.22pm, Friday, 14th May 2010]**
sketch_line (sketch_route is also affected) often produces erratic graphics. I have observed two issues:
1. apparently, if one of the stop members of the relation is an area tagged as amenity=bus_station, it is omitted completely. For example:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=80&style=padua]
consisting of relations:
[http://www.openstreetmap.org/browse/relation/611446]
[http://www.openstreetmap.org/browse/relation/611538]
The first stop (Bonola) is tagged as described above and missing from the sketch. This seems incorrect to me as, to my knowledge, bus_station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms.
2. Trolleybus lines (route=trolleybus) are not rendered at all - attempting to sketch them results in "No relation found".
To reproduce:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=92&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/913880]
[http://www.openstreetmap.org/browse/relation/913907]
3. in some cases incorrect stop names are displayed - names are mixed up or the same stop name is displayed multiple times. The graph itself seems correct (though I have not verified this in too much detail).
Mild example:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=19&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/661028]
[http://www.openstreetmap.org/browse/relation/661000]
Here just one stop name is wrong: the 11th (from Roserio) is displayed as "Duomo" when in fact it should be "Piazzale Accursio". (Duomo is the last stop.) Stops at which only one direction stops seem to be particular prone to this error (though it happens to all).
A more extreme one would be:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=78&style=padua]
relations:
[http://www.openstreetmap.org/browse/relation/420649]
[http://www.openstreetmap.org/browse/relation/445904]
This one is botched beyond recognition and usability: names are either missing completely, or the same name is repeated over and over multiple times.
And finally:
[http://78.46.81.38/api/sketch-line?network=ATM&ref=69&style=padua]
two alternative routes, resulting in a total of four relations:
[http://www.openstreetmap.org/browse/relation/536964]
[http://www.openstreetmap.org/browse/relation/537040]
[http://www.openstreetmap.org/browse/relation/536992]
[http://www.openstreetmap.org/browse/relation/537008] | defect | osm server side script sketch line and sketch route produce erratic output sketch line sketch route is also affected often produces erratic graphics i have observed two issues apparently if one of the stop members of the relation is an area tagged as amenity bus station it is omitted completely for example consisting of relations the first stop bonola is tagged as described above and missing from the sketch this seems incorrect to me as to my knowledge bus station is the tag to be used for a larger bus terminal where multiple lines stop at dedicated platforms trolleybus lines route trolleybus are not rendered at all attempting to sketch them results in no relation found to reproduce relations in some cases incorrect stop names are displayed names are mixed up or the same stop name is displayed multiple times the graph itself seems correct though i have not verified this in too much detail mild example relations here just one stop name is wrong the from roserio is displayed as duomo when in fact it should be piazzale accursio duomo is the last stop stops at which only one direction stops seem to be particular prone to this error though it happens to all a more extreme one would be relations this one is botched beyond recognition and usability names are either missing completely or the same name is repeated over and over multiple times and finally two alternative routes resulting in a total of four relations | 1 |
66,076 | 19,914,757,229 | IssuesEvent | 2022-01-25 21:10:59 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Space panel scroll gradient only appears after scrolling | T-Defect | ### Steps to reproduce
1. Be in lots of spaces
2. Open the app
3. Note lack of gradient
4. Scroll
5. Note gradient appears
6. Scroll all the way to the top (similar to app start)
7. Note gradient is still present
### Outcome
#### What did you expect?
The gradient should be consistently shown.
This gradient:

### Operating system
Windows 10
### Application version
Nightly (2022-01-25)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No | 1.0 | Space panel scroll gradient only appears after scrolling - ### Steps to reproduce
1. Be in lots of spaces
2. Open the app
3. Note lack of gradient
4. Scroll
5. Note gradient appears
6. Scroll all the way to the top (similar to app start)
7. Note gradient is still present
### Outcome
#### What did you expect?
The gradient should be consistently shown.
This gradient:

### Operating system
Windows 10
### Application version
Nightly (2022-01-25)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No | defect | space panel scroll gradient only appears after scrolling steps to reproduce be in lots of spaces open the app note lack of gradient scroll note gradient appears scroll all the way to the top similar to app start note gradient is still present outcome what did you expect the gradient should be consistently shown this gradient operating system windows application version nightly how did you install the app the internet homeserver io will you send logs no | 1 |
28,764 | 5,352,638,032 | IssuesEvent | 2017-02-20 00:19:07 | AtlasOfLivingAustralia/biocache-service | https://api.github.com/repos/AtlasOfLivingAustralia/biocache-service | closed | Download documentation - needs an edit | Downloads priority-low status-new type-defect | _From @nickdos on January 20, 2015 2:50_
_From @mbohun on August 19, 2014 11:16_
_migrated from:_ https://code.google.com/p/ala/issues/detail?id=230
_date:_ Mon Aug 12 03:24:21 2013
_author:_ johntan...@gmail.com
---
Records downloaded through the BioCache come with a README file which contains a link to information about the Download Fields:
[https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0](https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0)
Some of the fields have explanations that are not helpful. For example:
Coordinate Precision The precision of the coordinates
Coordinate Uncertainty In Meters Distance from the occurrence
Although there is a link to the Darwin Core Terms for the fields, the description could be more helpful for a person needing to understand what they have downloaded.
Other editing requirements
ACRONYMS need explaining eg GUID, IBRA, IMCRA
Spelling
Links to further assistance
Some issue flags missing definitions
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#39_
_Copied from original issue: AtlasOfLivingAustralia/bie-service#2_
| 1.0 | Download documentation - needs an edit - _From @nickdos on January 20, 2015 2:50_
_From @mbohun on August 19, 2014 11:16_
_migrated from:_ https://code.google.com/p/ala/issues/detail?id=230
_date:_ Mon Aug 12 03:24:21 2013
_author:_ johntan...@gmail.com
---
Records downloaded through the BioCache come with a README file which contains a link to information about the Download Fields:
[https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0](https://docs.google.com/spreadsheet/ccc?key=0AjNtzhUIIHeNdHhtcFVSM09qZ3c3N3ItUnBBc09TbHc#gid=0)
Some of the fields have explanations that are not helpful. For example:
Coordinate Precision The precision of the coordinates
Coordinate Uncertainty In Meters Distance from the occurrence
Although there is a link to the Darwin Core Terms for the fields, the description could be more helpful for a person needing to understand what they have downloaded.
Other editing requirements
ACRONYMS need explaining eg GUID, IBRA, IMCRA
Spelling
Links to further assistance
Some issue flags missing definitions
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#39_
_Copied from original issue: AtlasOfLivingAustralia/bie-service#2_
| defect | download documentation needs an edit from nickdos on january from mbohun on august migrated from date mon aug author johntan gmail com records downloaded through the biocache come with a readme file which contains a link to information about the download fields some of the fields have explanations that are not helpful for example coordinate precision the precision of the coordinates coordinate uncertainty in meters distance from the occurrence although there is a link to the darwin core terms for the fields the description could be more helpful for a person needing to understand what they have downloaded other editing requirements acronyms need explaining eg guid ibra imcra spelling links to further assistance some issue flags missing definitions copied from original issue atlasoflivingaustralia biocache hubs copied from original issue atlasoflivingaustralia bie service | 1 |
72,433 | 24,116,397,240 | IssuesEvent | 2022-09-20 15:01:40 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | objects sometimes create an unneeded L1 block | Type: Defect | ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | nixos
Distribution Version | 22.05
Kernel Version | 5.10.81
Architecture | x86-64
OpenZFS Version | 2.1.1-1
### Describe the problem you're observing
some objects are using 2 L0 block pointers directly in the dnode with no bonus data
other objects are using 1 L1 pointer in the dnode with 176 bytes of bonus data, and the L1 then contains 2 L0's, but the 2 L0's + bonus could have all fit within the dnode
https://github.com/openzfs/zfs/blob/master/include/sys/dnode.h#L241-L254
i think the root of the problem, is that this doesnt describe how to layout a dnode with 2 pointers, when the bonus is small enough to allow it
an example object with 2 L0's, no bonus, and no L1
```
Object lvl iblk dblk dsize dnsize lsize %full type
1 1 128K 16K 18K 512 32K 100.00 object directory (K=inherit) (Z=inherit=on)
dnode flags: USED_BYTES
dnode maxblkid: 1
Fat ZAP stats:
Indirect blocks:
0 L0 DVA[0]=<0:4800:800> DVA[1]=<0:2004800:800> DVA[2]=<0:10008800:800> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/800P birth=4L/4P fill=1 cksum=12f05825
4000 L0 DVA[0]=<0:5000:1000> DVA[1]=<0:2005000:1000> DVA[2]=<0:10009000:1000> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/1000P birth=4L/4P fill=1 cksum=9e86
```
an example object with a fatzap, the `.pack` file was 1 byte over the size limit of a microzap
it is stored as an L1 in the dnode, with 2 L0's, and 176 bytes of bouns
but the dnode could have held both L0's
```
Object lvl iblk dblk dsize dnsize lsize %full type
38 2 128K 16K 16K 512 32K 100.00 ZFS directory (K=inherit) (Z=inherit=uncompressed)
176 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED
dnode maxblkid: 1
path /lk/.git/objects/pack
uid 1000
gid 100
...
Fat ZAP stats:
...
pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.idx = 41 (type: Regular File)
pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.pack = 256 (type: Regular File)
Indirect blocks:
0 L1 DVA[0]=<0:42cf000:800> DVA[1]=<0:6015800:800> [L1 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=20000L/800P birth=20L/20P fill=2 cksum=d9559b8fb0ee03b5:819a23eca9f5b5
0 L0 DVA[0]=<0:818e000:800> DVA[1]=<0:a01a000:800> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/800P birth=20L/20P fill=1 cksum=5479b2c3a5a78dd8:3a4dad7bd5e005b
4000 L0 DVA[0]=<0:818e800:1000> DVA[1]=<0:a000000:1000> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/1000P birth=20L/20P fill=1 cksum=d37a5477b8ae53c4:343b2c00b22a
```
### Describe how to reproduce the problem
this test pool was just the result of running git clone on https://github.com/littlekernel/lk but it should be possible to reproduce it on any basic file with 2 records
| 1.0 | objects sometimes create an unneeded L1 block - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | nixos
Distribution Version | 22.05
Kernel Version | 5.10.81
Architecture | x86-64
OpenZFS Version | 2.1.1-1
### Describe the problem you're observing
some objects are using 2 L0 block pointers directly in the dnode with no bonus data
other objects are using 1 L1 pointer in the dnode with 176 bytes of bonus data, and the L1 then contains 2 L0's, but the 2 L0's + bonus could have all fit within the dnode
https://github.com/openzfs/zfs/blob/master/include/sys/dnode.h#L241-L254
i think the root of the problem, is that this doesnt describe how to layout a dnode with 2 pointers, when the bonus is small enough to allow it
an example object with 2 L0's, no bonus, and no L1
```
Object lvl iblk dblk dsize dnsize lsize %full type
1 1 128K 16K 18K 512 32K 100.00 object directory (K=inherit) (Z=inherit=on)
dnode flags: USED_BYTES
dnode maxblkid: 1
Fat ZAP stats:
Indirect blocks:
0 L0 DVA[0]=<0:4800:800> DVA[1]=<0:2004800:800> DVA[2]=<0:10008800:800> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/800P birth=4L/4P fill=1 cksum=12f05825
4000 L0 DVA[0]=<0:5000:1000> DVA[1]=<0:2005000:1000> DVA[2]=<0:10009000:1000> [L0 object directory] fletcher4 lz4 unencrypted LE contiguous unique triple size=4000L/1000P birth=4L/4P fill=1 cksum=9e86
```
an example object with a fatzap, the `.pack` file was 1 byte over the size limit of a microzap
it is stored as an L1 in the dnode, with 2 L0's, and 176 bytes of bouns
but the dnode could have held both L0's
```
Object lvl iblk dblk dsize dnsize lsize %full type
38 2 128K 16K 16K 512 32K 100.00 ZFS directory (K=inherit) (Z=inherit=uncompressed)
176 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED
dnode maxblkid: 1
path /lk/.git/objects/pack
uid 1000
gid 100
...
Fat ZAP stats:
...
pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.idx = 41 (type: Regular File)
pack-66d8736892f64aa480580a9c5a081a1f6cbabb90.pack = 256 (type: Regular File)
Indirect blocks:
0 L1 DVA[0]=<0:42cf000:800> DVA[1]=<0:6015800:800> [L1 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=20000L/800P birth=20L/20P fill=2 cksum=d9559b8fb0ee03b5:819a23eca9f5b5
0 L0 DVA[0]=<0:818e000:800> DVA[1]=<0:a01a000:800> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/800P birth=20L/20P fill=1 cksum=5479b2c3a5a78dd8:3a4dad7bd5e005b
4000 L0 DVA[0]=<0:818e800:1000> DVA[1]=<0:a000000:1000> [L0 ZFS directory] sha256 lz4 unencrypted LE contiguous unique double size=4000L/1000P birth=20L/20P fill=1 cksum=d37a5477b8ae53c4:343b2c00b22a
```
### Describe how to reproduce the problem
this test pool was just the result of running git clone on https://github.com/littlekernel/lk but it should be possible to reproduce it on any basic file with 2 records
| defect | objects sometimes create an unneeded block system information type version name distribution name nixos distribution version kernel version architecture openzfs version describe the problem you re observing some objects are using block pointers directly in the dnode with no bonus data other objects are using pointer in the dnode with bytes of bonus data and the then contains s but the s bonus could have all fit within the dnode i think the root of the problem is that this doesnt describe how to layout a dnode with pointers when the bonus is small enough to allow it an example object with s no bonus and no object lvl iblk dblk dsize dnsize lsize full type object directory k inherit z inherit on dnode flags used bytes dnode maxblkid fat zap stats indirect blocks dva dva dva unencrypted le contiguous unique triple size birth fill cksum dva dva dva unencrypted le contiguous unique triple size birth fill cksum an example object with a fatzap the pack file was byte over the size limit of a microzap it is stored as an in the dnode with s and bytes of bouns but the dnode could have held both s object lvl iblk dblk dsize dnsize lsize full type zfs directory k inherit z inherit uncompressed bonus system attributes dnode flags used bytes userused accounted userobjused accounted dnode maxblkid path lk git objects pack uid gid fat zap stats pack idx type regular file pack pack type regular file indirect blocks dva dva unencrypted le contiguous unique double size birth fill cksum dva dva unencrypted le contiguous unique double size birth fill cksum dva dva unencrypted le contiguous unique double size birth fill cksum describe how to reproduce the problem this test pool was just the result of running git clone on but it should be possible to reproduce it on any basic file with records | 1 |
187,545 | 22,045,797,556 | IssuesEvent | 2022-05-30 01:27:02 | utopikkad/my-Todo-List | https://api.github.com/repos/utopikkad/my-Todo-List | closed | CVE-2021-37701 (High) detected in tar-4.4.13.tgz, tar-4.4.1.tgz - autoclosed | security vulnerability | ## CVE-2021-37701 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b></p></summary>
<p>
<details><summary><b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>
Dependency Hierarchy:
- cli-7.3.9.tgz (Root Library)
- pacote-9.4.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p>
<p>
Dependency Hierarchy:
- karma-3.1.4.tgz (Root Library)
- chokidar-2.0.4.tgz
- fsevents-1.2.4.tgz
- node-pre-gyp-0.10.0.tgz
- :x: **tar-4.4.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/utopikkad/my-Todo-List/commit/bd19af551d149e2bce7801819589f1a69368d63a">bd19af551d149e2bce7801819589f1a69368d63a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.16</p>
<p>Direct dependency fix Resolution (@angular/cli): 7.3.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-37701 (High) detected in tar-4.4.13.tgz, tar-4.4.1.tgz - autoclosed - ## CVE-2021-37701 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.1.tgz</b></p></summary>
<p>
<details><summary><b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>
Dependency Hierarchy:
- cli-7.3.9.tgz (Root Library)
- pacote-9.4.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
</details>
<details><summary><b>tar-4.4.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p>
<p>
Dependency Hierarchy:
- karma-3.1.4.tgz (Root Library)
- chokidar-2.0.4.tgz
- fsevents-1.2.4.tgz
- node-pre-gyp-0.10.0.tgz
- :x: **tar-4.4.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/utopikkad/my-Todo-List/commit/bd19af551d149e2bce7801819589f1a69368d63a">bd19af551d149e2bce7801819589f1a69368d63a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.16</p>
<p>Direct dependency fix Resolution (@angular/cli): 7.3.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in tar tgz tar tgz autoclosed cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar tgz tar for node library home page a href dependency hierarchy cli tgz root library pacote tgz x tar tgz vulnerable library tar tgz tar for node library home page a href dependency hierarchy karma tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution angular cli step up your open source security game with whitesource | 0 |
78,172 | 27,355,813,306 | IssuesEvent | 2023-02-27 12:48:04 | matrix-org/synapse | https://api.github.com/repos/matrix-org/synapse | opened | Initial syncs being cached means you can get an outdated view of account data | A-Sync S-Minor T-Defect O-Uncommon | There are two problems with this:
- (1) it possibly betrays client expectations, that doing an initial sync will get them to 'correct' state immediately.
- (2) it means we don't apply the correct set of ignored users to the events received in the timeline of the initial sync. (For context: the server hides events sent by ignored users, when sending events to the client.)
This issue is causing https://github.com/vector-im/element-ios/issues/7387 — we *could* tell them to just do a zero-timeout incremental sync immediately but that feels like a hack and it wouldn't solve (2).
For (2), I wonder if we can just fetch the latest account data stream ID for the ignored users list, when doing an initial sync, and embedding it into the sync response cache token.
That way, the sync response cache entry will automatically be invalidated when the account data changes.
It's not a perfect solution since it only helps with ignored users, but it at least solves (2).
Problem (1) might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data. Sliding sync might obviate these concerns anyway. | 1.0 | Initial syncs being cached means you can get an outdated view of account data - There are two problems with this:
- (1) it possibly betrays client expectations, that doing an initial sync will get them to 'correct' state immediately.
- (2) it means we don't apply the correct set of ignored users to the events received in the timeline of the initial sync. (For context: the server hides events sent by ignored users, when sending events to the client.)
This issue is causing https://github.com/vector-im/element-ios/issues/7387 — we *could* tell them to just do a zero-timeout incremental sync immediately but that feels like a hack and it wouldn't solve (2).
For (2), I wonder if we can just fetch the latest account data stream ID for the ignored users list, when doing an initial sync, and embedding it into the sync response cache token.
That way, the sync response cache entry will automatically be invalidated when the account data changes.
It's not a perfect solution since it only helps with ignored users, but it at least solves (2).
Problem (1) might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data. Sliding sync might obviate these concerns anyway. | defect | initial syncs being cached means you can get an outdated view of account data there are two problems with this it possibly betrays client expectations that doing an initial sync will get them to correct state immediately it means we don t apply the correct set of ignored users to the events received in the timeline of the initial sync for context the server hides events sent by ignored users when sending events to the client this issue is causing — we could tell them to just do a zero timeout incremental sync immediately but that feels like a hack and it wouldn t solve for i wonder if we can just fetch the latest account data stream id for the ignored users list when doing an initial sync and embedding it into the sync response cache token that way the sync response cache entry will automatically be invalidated when the account data changes it s not a perfect solution since it only helps with ignored users but it at least solves problem might be solvable through better documentation about the fact that clients should always incrementally sync after an initial sync if they want to guarantee they have the freshest data sliding sync might obviate these concerns anyway | 1 |
814,247 | 30,496,367,288 | IssuesEvent | 2023-07-18 11:07:14 | ufosc/Club_Website_2 | https://api.github.com/repos/ufosc/Club_Website_2 | closed | V0.2 BETA | help wanted discussion high priority | ## V0.2 Beta Release Issue Tracker
| Issue | Ref | Status | Assignee |
|-------|-----|--------|----------|
| TEST: /api/edit ROUTE | #158 | **HELP WANTED** | Thinura |
| ~TEST: /api/blog route~ | #157 | DONE | @zeim839 |
| ~TEST: /api/image route~ | N/A | DONE | @zeim839 |
Goals: Finish testing before releasing beta to public. | 1.0 | V0.2 BETA - ## V0.2 Beta Release Issue Tracker
| Issue | Ref | Status | Assignee |
|-------|-----|--------|----------|
| TEST: /api/edit ROUTE | #158 | **HELP WANTED** | Thinura |
| ~TEST: /api/blog route~ | #157 | DONE | @zeim839 |
| ~TEST: /api/image route~ | N/A | DONE | @zeim839 |
Goals: Finish testing before releasing beta to public. | non_defect | beta beta release issue tracker issue ref status assignee test api edit route help wanted thinura test api blog route done test api image route n a done goals finish testing before releasing beta to public | 0 |
31,799 | 6,626,437,349 | IssuesEvent | 2017-09-22 19:33:15 | wheat32/Armold | https://api.github.com/repos/wheat32/Armold | closed | Close sensors at end of runtime | defect medium priority | Add a method to close all the sensors at the end of runtime. This can be done by calling `Sensor.close();`. | 1.0 | Close sensors at end of runtime - Add a method to close all the sensors at the end of runtime. This can be done by calling `Sensor.close();`. | defect | close sensors at end of runtime add a method to close all the sensors at the end of runtime this can be done by calling sensor close | 1 |
158,594 | 12,419,653,383 | IssuesEvent | 2020-05-23 07:34:10 | the-canonizer/canonizer.2.0 | https://api.github.com/repos/the-canonizer/canonizer.2.0 | closed | My Supports : Camp names are missing and getting page crash on remove support of the blank camp name | 18th release Fixed Show Stopper bug ready to test | Steps to reproduce :
1. Go to https://staging.canonizer.com/
2. Login with valid email and password
3. Go to browse
4. Select Namespace as "Sandbox testing" and select as of filter "as of date" = "6/18/2019, 5:30:00 AM"
5. Go to Account settings -> My Supports
Please refer attached screenshot

I think filter values are getting saved in session.User supports page also not listing all the supported camps
Please confirm..
| 1.0 | My Supports : Camp names are missing and getting page crash on remove support of the blank camp name - Steps to reproduce :
1. Go to https://staging.canonizer.com/
2. Login with valid email and password
3. Go to browse
4. Select Namespace as "Sandbox testing" and select as of filter "as of date" = "6/18/2019, 5:30:00 AM"
5. Go to Account settings -> My Supports
Please refer attached screenshot

I think filter values are getting saved in session.User supports page also not listing all the supported camps
Please confirm..
| non_defect | my supports camp names are missing and getting page crash on remove support of the blank camp name steps to reproduce go to login with valid email and password go to browse select namespace as sandbox testing and select as of filter as of date am go to account settings my supports please refer attached screenshot i think filter values are getting saved in session user supports page also not listing all the supported camps please confirm | 0 |
146,551 | 19,406,174,645 | IssuesEvent | 2021-12-20 01:14:18 | keanhankins/ranger | https://api.github.com/repos/keanhankins/ranger | opened | CVE-2021-4104 (High) detected in log4j-1.2.17.jar | security vulnerability | ## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: ranger/ranger-hive-plugin-shim/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/ranger-hive-plugin-shim/pom.xml","/ranger-examples/plugin-sampleapp/pom.xml","/plugin-nifi/pom.xml","/ranger-kms-plugin-shim/pom.xml","/ranger-kylin-plugin-shim/pom.xml","/intg/pom.xml","/ranger-hbase-plugin-shim/pom.xml","/agents-cred/pom.xml","/ranger-knox-plugin-shim/pom.xml","/hdfs-agent/pom.xml","/ranger-examples/sampleapp/pom.xml","/agents-audit/pom.xml","/agents-common/pom.xml","/security-admin/pom.xml","/ranger-presto-plugin-shim/pom.xml","/ranger-atlas-plugin-shim/pom.xml","/ranger-examples/conditions-enrichers/pom.xml","/plugin-presto/pom.xml","/ranger-ozone-plugin-shim/pom.xml","/plugin-ozone/pom.xml","/ranger-kafka-plugin-shim/pom.xml","/distro/pom.xml","/ugsync/pom.xml","/ranger-tools/pom.xml","/plugin-nifi-registry/pom.xml","/kms/pom.xml","/plugin-atlas/pom.xml","/plugin-kms/pom.xml","/plugin-kylin/pom.xml","/plugin-yarn/pom.xml","/plugin-kudu/pom.xml","/plugin-solr/pom.xml","/tagsync/pom.xml","/ranger-elasticsearch-plugin-shim/pom.xml","/plugin-elasticsearch/pom.xml","/storm-agent/pom.xml","/ranger-solr-plugin-shim/pom.xml","/ranger-hdfs-plugin-shim/pom.xml","/unixauthclient/pom.xml","/hbase-agent/pom.xml","/ranger-storm-plugin-shim/pom.xml","/ranger-yarn-plugin-shim/pom.xml","/hive-agent/pom.xml","/plugin-kafka/pom.xml","/credentialbuilder/pom.xml","/embeddedwebserver/pom.xml","/ranger-examples/sample-client/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-4104 (High) detected in log4j-1.2.17.jar - ## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: ranger/ranger-hive-plugin-shim/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/ranger-hive-plugin-shim/pom.xml","/ranger-examples/plugin-sampleapp/pom.xml","/plugin-nifi/pom.xml","/ranger-kms-plugin-shim/pom.xml","/ranger-kylin-plugin-shim/pom.xml","/intg/pom.xml","/ranger-hbase-plugin-shim/pom.xml","/agents-cred/pom.xml","/ranger-knox-plugin-shim/pom.xml","/hdfs-agent/pom.xml","/ranger-examples/sampleapp/pom.xml","/agents-audit/pom.xml","/agents-common/pom.xml","/security-admin/pom.xml","/ranger-presto-plugin-shim/pom.xml","/ranger-atlas-plugin-shim/pom.xml","/ranger-examples/conditions-enrichers/pom.xml","/plugin-presto/pom.xml","/ranger-ozone-plugin-shim/pom.xml","/plugin-ozone/pom.xml","/ranger-kafka-plugin-shim/pom.xml","/distro/pom.xml","/ugsync/pom.xml","/ranger-tools/pom.xml","/plugin-nifi-registry/pom.xml","/kms/pom.xml","/plugin-atlas/pom.xml","/plugin-kms/pom.xml","/plugin-kylin/pom.xml","/plugin-yarn/pom.xml","/plugin-kudu/pom.xml","/plugin-solr/pom.xml","/tagsync/pom.xml","/ranger-elasticsearch-plugin-shim/pom.xml","/plugin-elasticsearch/pom.xml","/storm-agent/pom.xml","/ranger-solr-plugin-shim/pom.xml","/ranger-hdfs-plugin-shim/pom.xml","/unixauthclient/pom.xml","/hbase-agent/pom.xml","/ranger-storm-plugin-shim/pom.xml","/ranger-yarn-plugin-shim/pom.xml","/hive-agent/pom.xml","/plugin-kafka/pom.xml","/credentialbuilder/pom.xml","/embeddedwebserver/pom.xml","/ranger-examples/sample-client/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-4104","vulnerabilityDetails":"JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-4104","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file ranger ranger hive plugin shim pom xml path to vulnerable library home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar canner repository jar canner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar canner repository jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in base branch master vulnerability details jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions vulnerabilityurl | 0 |
202,171 | 23,061,486,630 | IssuesEvent | 2022-07-25 10:19:59 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 106: envoy-1.17.3: 4 advisories [8.3] | 1.severity: security | [search](https://search.nix.gsc.io/?q=envoy&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=envoy+in%3Apath&type=Code)
* [ ] [CVE-2021-32777](https://nvd.nist.gov/vuln/detail/CVE-2021-32777) CVSSv3=8.3 (nixos-unstable)
* [ ] [CVE-2021-32779](https://nvd.nist.gov/vuln/detail/CVE-2021-32779) CVSSv3=8.3 (nixos-unstable)
* [ ] [CVE-2021-32778](https://nvd.nist.gov/vuln/detail/CVE-2021-32778) CVSSv3=7.5 (nixos-unstable)
* [ ] [CVE-2021-32781](https://nvd.nist.gov/vuln/detail/CVE-2021-32781) CVSSv3=7.5 (nixos-unstable)
## CVE details
### CVE-2021-32777
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions when ext-authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the HTTP spec. However, only the last header value is sent. This may allow specifically crafted requests to bypass authorization. Attackers may be able to escalate privileges when using ext-authz extension or back end service that uses multiple value headers for authorization. A specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext-authz extension. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to the ext-authz extension to correctly merge multiple request header values, when sending request for authorization.
### CVE-2021-32779
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy incorrectly handled a URI '#fragment' element as part of the path element. Envoy is configured with an RBAC filter for authorization or similar mechanism with an explicit case of a final "/admin" path element, or is using a negative assertion with final path element of "/admin". The client sends request to "/app1/admin#foo". In Envoy prior to 1.18.0, or 1.18.0+ configured with path_normalization=false. Envoy treats fragment as a suffix of the query string when present, or as a suffix of the path when query string is absent, so it evaluates the final path element as "/admin#foo" and mismatches with the configured "/admin" path element. In Envoy 1.18.0+ configured with path_normalization=true. Envoy transforms this to /app1/admin%23foo and mismatches with the configured /admin prefix. The resulting URI is sent to the next server-agent with the offending "#foo" fragment which violates RFC3986 or with the nonsensical "%23foo" text appended. A specifically constructed request with URI containing '#fragment' element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of Privileges when path based request authorization extensions. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes that removes fragment from URI path in incoming requests.
### CVE-2021-32778
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy’s procedure for resetting a HTTP/2 stream has O(N^2) complexity, leading to high CPU utilization when a large number of streams are reset. Deployments are susceptible to Denial of Service when Envoy is configured with high limit on H/2 concurrent streams. An attacker wishing to exploit this vulnerability would require a client opening and closing a large number of H/2 streams. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to reduce time complexity of resetting HTTP/2 streams. As a workaround users may limit the number of simultaneous HTTP/2 dreams for upstream and downstream peers to a low number, i.e. 100.
### CVE-2021-32781
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions after Envoy sends a locally generated response it must stop further processing of request or response data. However when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block. A specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a Denial of Service when using extensions that modify and increase the size of request or response bodies, such as decompressor filter. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to address incomplete termination of request processing after locally generated response. As a workaround disable Envoy's decompressor, json-transcoder or grpc-web extensions or proprietary extensions that modify and increase the size of request or response bodies, if feasible.
-----
Scanned versions: nixos-unstable: bc9b956714e.
Cc @lukegb
| True | Vulnerability roundup 106: envoy-1.17.3: 4 advisories [8.3] - [search](https://search.nix.gsc.io/?q=envoy&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=envoy+in%3Apath&type=Code)
* [ ] [CVE-2021-32777](https://nvd.nist.gov/vuln/detail/CVE-2021-32777) CVSSv3=8.3 (nixos-unstable)
* [ ] [CVE-2021-32779](https://nvd.nist.gov/vuln/detail/CVE-2021-32779) CVSSv3=8.3 (nixos-unstable)
* [ ] [CVE-2021-32778](https://nvd.nist.gov/vuln/detail/CVE-2021-32778) CVSSv3=7.5 (nixos-unstable)
* [ ] [CVE-2021-32781](https://nvd.nist.gov/vuln/detail/CVE-2021-32781) CVSSv3=7.5 (nixos-unstable)
## CVE details
### CVE-2021-32777
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions when ext-authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the HTTP spec. However, only the last header value is sent. This may allow specifically crafted requests to bypass authorization. Attackers may be able to escalate privileges when using ext-authz extension or back end service that uses multiple value headers for authorization. A specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext-authz extension. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to the ext-authz extension to correctly merge multiple request header values, when sending request for authorization.
### CVE-2021-32779
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy incorrectly handled a URI '#fragment' element as part of the path element. Envoy is configured with an RBAC filter for authorization or similar mechanism with an explicit case of a final "/admin" path element, or is using a negative assertion with final path element of "/admin". The client sends request to "/app1/admin#foo". In Envoy prior to 1.18.0, or 1.18.0+ configured with path_normalization=false. Envoy treats fragment as a suffix of the query string when present, or as a suffix of the path when query string is absent, so it evaluates the final path element as "/admin#foo" and mismatches with the configured "/admin" path element. In Envoy 1.18.0+ configured with path_normalization=true. Envoy transforms this to /app1/admin%23foo and mismatches with the configured /admin prefix. The resulting URI is sent to the next server-agent with the offending "#foo" fragment which violates RFC3986 or with the nonsensical "%23foo" text appended. A specifically constructed request with URI containing '#fragment' element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of Privileges when path based request authorization extensions. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes that removes fragment from URI path in incoming requests.
### CVE-2021-32778
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions envoy’s procedure for resetting a HTTP/2 stream has O(N^2) complexity, leading to high CPU utilization when a large number of streams are reset. Deployments are susceptible to Denial of Service when Envoy is configured with high limit on H/2 concurrent streams. An attacker wishing to exploit this vulnerability would require a client opening and closing a large number of H/2 streams. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to reduce time complexity of resetting HTTP/2 streams. As a workaround users may limit the number of simultaneous HTTP/2 dreams for upstream and downstream peers to a low number, i.e. 100.
### CVE-2021-32781
Envoy is an open source L7 proxy and communication bus designed for large modern service oriented architectures. In affected versions after Envoy sends a locally generated response it must stop further processing of request or response data. However when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block. A specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a Denial of Service when using extensions that modify and increase the size of request or response bodies, such as decompressor filter. Envoy versions 1.19.1, 1.18.4, 1.17.4, 1.16.5 contain fixes to address incomplete termination of request processing after locally generated response. As a workaround disable Envoy's decompressor, json-transcoder or grpc-web extensions or proprietary extensions that modify and increase the size of request or response bodies, if feasible.
-----
Scanned versions: nixos-unstable: bc9b956714e.
Cc @lukegb
| non_defect | vulnerability roundup envoy advisories nixos unstable nixos unstable nixos unstable nixos unstable cve details cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions when ext authz extension is sending request headers to the external authorization service it must merge multiple value headers according to the http spec however only the last header value is sent this may allow specifically crafted requests to bypass authorization attackers may be able to escalate privileges when using ext authz extension or back end service that uses multiple value headers for authorization a specifically constructed request may be delivered by an untrusted downstream peer in the presence of ext authz extension envoy versions contain fixes to the ext authz extension to correctly merge multiple request header values when sending request for authorization cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions envoy incorrectly handled a uri fragment element as part of the path element envoy is configured with an rbac filter for authorization or similar mechanism with an explicit case of a final admin path element or is using a negative assertion with final path element of admin the client sends request to admin foo in envoy prior to or configured with path normalization false envoy treats fragment as a suffix of the query string when present or as a suffix of the path when query string is absent so it evaluates the final path element as admin foo and mismatches with the configured admin path element in envoy configured with path normalization true envoy transforms this to admin and mismatches with the configured admin prefix the resulting uri is sent to the next server agent with the offending foo fragment which violates or with the nonsensical text appended a specifically constructed request with uri containing fragment element delivered by an untrusted client in the presence of path based request authorization resulting in escalation of privileges when path based request authorization extensions envoy versions contain fixes that removes fragment from uri path in incoming requests cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions envoy’s procedure for resetting a http stream has o n complexity leading to high cpu utilization when a large number of streams are reset deployments are susceptible to denial of service when envoy is configured with high limit on h concurrent streams an attacker wishing to exploit this vulnerability would require a client opening and closing a large number of h streams envoy versions contain fixes to reduce time complexity of resetting http streams as a workaround users may limit the number of simultaneous http dreams for upstream and downstream peers to a low number i e cve envoy is an open source proxy and communication bus designed for large modern service oriented architectures in affected versions after envoy sends a locally generated response it must stop further processing of request or response data however when local response is generated due the internal buffer overflow while request or response is processed by the filter chain the operation may not be stopped completely and result in accessing a freed memory block a specifically constructed request delivered by an untrusted downstream or upstream peer in the presence of extensions that modify and increase the size of request or response bodies resulting in a denial of service when using extensions that modify and increase the size of request or response bodies such as decompressor filter envoy versions contain fixes to address incomplete termination of request processing after locally generated response as a workaround disable envoy s decompressor json transcoder or grpc web extensions or proprietary extensions that modify and increase the size of request or response bodies if feasible scanned versions nixos unstable cc lukegb | 0 |
27,930 | 5,412,428,396 | IssuesEvent | 2017-03-01 14:32:27 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | Unespected behaviour drawing arc with ellipse() | bug category: documentation | #### System information (version)
- OpenCV => 3.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2015
#### Detailed description
Both doc and code works different than expected. It looks that `startAngle` and `endAngle` refer to the projections on the ellipse of the wanted arc on the bounding circle. Look at the image

The blue arc has been generated 30/45/135 as angle/start/end.
The red arc has been using same angles as above but choosing equal axis.
May be this is wanted but is not same as is in [the doc](http://docs.opencv.org/3.2.0/d6/d6e/group__imgproc__draw.html#ga28b2267d35786f5f890ca167236cbc69).
#### Steps to reproduce
```.cpp
int TestEllipse()
{
int angle = 30,startAngle = 45,endAngle = 135;
Mat img(400, 400, CV_8UC3, Scalar(255, 255, 255));
Point center = img.size() / 2;
int major = 0.4*img.rows;
Size sz(major, major*0.5);
line(img, Point(center.x, 0), Point(center.x, img.rows), Scalar(0), 1);
line(img, Point(0, center.y), Point(img.cols, center.y), Scalar(0), 1);
ellipse(img, center, sz, angle, 0, 360, CL_GREEN, 1);
ellipse(img, center, sz, angle, startAngle, endAngle, CL_BLUE, 2);
sz = Size(sz.width, sz.width);
ellipse(img, center, sz, angle, 0, 360, CL_MAGENTA, 1);
ellipse(img, center, sz, angle, startAngle, endAngle, CL_RED, 2);
vector<double> mags, angles, x, y;
double mag = 1.2 * major, alpha = angle;
mags = { mag, mag };
angles = { (alpha + startAngle), (alpha + endAngle) };
polarToCart(mags, angles, x, y, true);
for (size_t i = 0; i < x.size(); i++)
line(img, center, center + Point(x[i], y[i]), Scalar(160, 160, 160));
imshow("ellipse", img);
waitKey(0);
return 0;
}
```
| 1.0 | Unespected behaviour drawing arc with ellipse() - #### System information (version)
- OpenCV => 3.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2015
#### Detailed description
Both doc and code works different than expected. It looks that `startAngle` and `endAngle` refer to the projections on the ellipse of the wanted arc on the bounding circle. Look at the image

The blue arc has been generated 30/45/135 as angle/start/end.
The red arc has been using same angles as above but choosing equal axis.
May be this is wanted but is not same as is in [the doc](http://docs.opencv.org/3.2.0/d6/d6e/group__imgproc__draw.html#ga28b2267d35786f5f890ca167236cbc69).
#### Steps to reproduce
```.cpp
int TestEllipse()
{
int angle = 30,startAngle = 45,endAngle = 135;
Mat img(400, 400, CV_8UC3, Scalar(255, 255, 255));
Point center = img.size() / 2;
int major = 0.4*img.rows;
Size sz(major, major*0.5);
line(img, Point(center.x, 0), Point(center.x, img.rows), Scalar(0), 1);
line(img, Point(0, center.y), Point(img.cols, center.y), Scalar(0), 1);
ellipse(img, center, sz, angle, 0, 360, CL_GREEN, 1);
ellipse(img, center, sz, angle, startAngle, endAngle, CL_BLUE, 2);
sz = Size(sz.width, sz.width);
ellipse(img, center, sz, angle, 0, 360, CL_MAGENTA, 1);
ellipse(img, center, sz, angle, startAngle, endAngle, CL_RED, 2);
vector<double> mags, angles, x, y;
double mag = 1.2 * major, alpha = angle;
mags = { mag, mag };
angles = { (alpha + startAngle), (alpha + endAngle) };
polarToCart(mags, angles, x, y, true);
for (size_t i = 0; i < x.size(); i++)
line(img, center, center + Point(x[i], y[i]), Scalar(160, 160, 160));
imshow("ellipse", img);
waitKey(0);
return 0;
}
```
| non_defect | unespected behaviour drawing arc with ellipse system information version opencv operating system platform windows bit compiler visual studio detailed description both doc and code works different than expected it looks that startangle and endangle refer to the projections on the ellipse of the wanted arc on the bounding circle look at the image the blue arc has been generated as angle start end the red arc has been using same angles as above but choosing equal axis may be this is wanted but is not same as is in steps to reproduce cpp int testellipse int angle startangle endangle mat img cv scalar point center img size int major img rows size sz major major line img point center x point center x img rows scalar line img point center y point img cols center y scalar ellipse img center sz angle cl green ellipse img center sz angle startangle endangle cl blue sz size sz width sz width ellipse img center sz angle cl magenta ellipse img center sz angle startangle endangle cl red vector mags angles x y double mag major alpha angle mags mag mag angles alpha startangle alpha endangle polartocart mags angles x y true for size t i i x size i line img center center point x y scalar imshow ellipse img waitkey return | 0 |
450,415 | 31,899,771,720 | IssuesEvent | 2023-09-18 06:53:21 | jacksonhooi/INF2001_P5-4 | https://api.github.com/repos/jacksonhooi/INF2001_P5-4 | closed | #9 Identify Non-Functional User Requirements | documentation | Task Description:
Discuss with the group and identify non-functional user requirements.
Initial Requirements (from project brief)
1. The app should be Web-based in a language of your choosing
2. The manager should be able to visualise the staff workload immediately on the landing page
3. The manager should be able to allocate jobs to staff for one week at a time
4. The manager should be able to view up to three staff availability and any relevant information
to make the job assignment easier on the job allocation page
5. When displaying the staff availability, the workload assigned, staff’s job preference, staff’s
location at a particular date, and availabilities for the week should be shown
6. On the manager’s landing page, the top three staff with the lowest workload should be shown,
and highlight all staff over 40 hours of jobs allocated
7. Staff should be able to view their weekly job assignments and overall workload for the month
on their landing page
8. Staff can add and edit their availabilities up to 5 weeks ahead of time.
9. Staff can indicate their job preference for the week
10. Staff can reject jobs assigned to them, but they will be warned to discuss the jobs with their
manager before proceeding with the rejection
11. The company’s IT administrators will oversee adding new staff and managers to the system
Goals:
- [x] List of non-functional requirements created
Completion Criteria
Successful creation of list of non functional requirements
Allocated Time:
10 Days
Start date: 5 Sept
End date: 15 Sept | 1.0 | #9 Identify Non-Functional User Requirements - Task Description:
Discuss with the group and identify non-functional user requirements.
Initial Requirements (from project brief)
1. The app should be Web-based in a language of your choosing
2. The manager should be able to visualise the staff workload immediately on the landing page
3. The manager should be able to allocate jobs to staff for one week at a time
4. The manager should be able to view up to three staff availability and any relevant information
to make the job assignment easier on the job allocation page
5. When displaying the staff availability, the workload assigned, staff’s job preference, staff’s
location at a particular date, and availabilities for the week should be shown
6. On the manager’s landing page, the top three staff with the lowest workload should be shown,
and highlight all staff over 40 hours of jobs allocated
7. Staff should be able to view their weekly job assignments and overall workload for the month
on their landing page
8. Staff can add and edit their availabilities up to 5 weeks ahead of time.
9. Staff can indicate their job preference for the week
10. Staff can reject jobs assigned to them, but they will be warned to discuss the jobs with their
manager before proceeding with the rejection
11. The company’s IT administrators will oversee adding new staff and managers to the system
Goals:
- [x] List of non-functional requirements created
Completion Criteria
Successful creation of list of non functional requirements
Allocated Time:
10 Days
Start date: 5 Sept
End date: 15 Sept | non_defect | identify non functional user requirements task description discuss with the group and identify non functional user requirements initial requirements from project brief the app should be web based in a language of your choosing the manager should be able to visualise the staff workload immediately on the landing page the manager should be able to allocate jobs to staff for one week at a time the manager should be able to view up to three staff availability and any relevant information to make the job assignment easier on the job allocation page when displaying the staff availability the workload assigned staff’s job preference staff’s location at a particular date and availabilities for the week should be shown on the manager’s landing page the top three staff with the lowest workload should be shown and highlight all staff over hours of jobs allocated staff should be able to view their weekly job assignments and overall workload for the month on their landing page staff can add and edit their availabilities up to weeks ahead of time staff can indicate their job preference for the week staff can reject jobs assigned to them but they will be warned to discuss the jobs with their manager before proceeding with the rejection the company’s it administrators will oversee adding new staff and managers to the system goals list of non functional requirements created completion criteria successful creation of list of non functional requirements allocated time days start date sept end date sept | 0 |
37,124 | 8,215,468,696 | IssuesEvent | 2018-09-05 05:41:04 | masteroy/algorithm | https://api.github.com/repos/masteroy/algorithm | closed | [LeetCode] 150. Evaluate Reverse Polish Notation | LeetCode Medium | Evaluate the value of an arithmetic expression in Reverse Polish Notation.
Valid operators are +, -, *, /. Each operand may be an integer or another expression.
**Note:**
- The division between two integers should truncate toward zero.
- The given RPN expression is always valid. That means the expression would always evaluate to a result and there won't be any divide by zero operation.
**Example 1:**
```
Input: ["2", "1", "+", "3", "*"]
Output: 9
Explanation: ((2 + 1) * 3) = 9
```
**Example 2:**
```
Input: ["4", "13", "5", "/", "+"]
Output: 6
Explanation: (4 + (13 / 5)) = 6
```
**Example 3:**
```
Input: ["10", "6", "9", "3", "+", "-11", "*", "/", "*", "17", "+", "5", "+"]
Output: 22
Explanation:
((10 * (6 / ((9 + 3) * -11))) + 17) + 5
= ((10 * (6 / (12 * -11))) + 17) + 5
= ((10 * (6 / -132)) + 17) + 5
= ((10 * 0) + 17) + 5
= (0 + 17) + 5
= 17 + 5
= 22
``` | 1.0 | [LeetCode] 150. Evaluate Reverse Polish Notation - Evaluate the value of an arithmetic expression in Reverse Polish Notation.
Valid operators are +, -, *, /. Each operand may be an integer or another expression.
**Note:**
- The division between two integers should truncate toward zero.
- The given RPN expression is always valid. That means the expression would always evaluate to a result and there won't be any divide by zero operation.
**Example 1:**
```
Input: ["2", "1", "+", "3", "*"]
Output: 9
Explanation: ((2 + 1) * 3) = 9
```
**Example 2:**
```
Input: ["4", "13", "5", "/", "+"]
Output: 6
Explanation: (4 + (13 / 5)) = 6
```
**Example 3:**
```
Input: ["10", "6", "9", "3", "+", "-11", "*", "/", "*", "17", "+", "5", "+"]
Output: 22
Explanation:
((10 * (6 / ((9 + 3) * -11))) + 17) + 5
= ((10 * (6 / (12 * -11))) + 17) + 5
= ((10 * (6 / -132)) + 17) + 5
= ((10 * 0) + 17) + 5
= (0 + 17) + 5
= 17 + 5
= 22
``` | non_defect | evaluate reverse polish notation evaluate the value of an arithmetic expression in reverse polish notation valid operators are each operand may be an integer or another expression note the division between two integers should truncate toward zero the given rpn expression is always valid that means the expression would always evaluate to a result and there won t be any divide by zero operation example input output explanation example input output explanation example input output explanation | 0 |
1,625 | 2,603,968,448 | IssuesEvent | 2015-02-24 18:59:39 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳阴茎里面有疙瘩 | auto-migrated Priority-Medium Type-Defect | ```
沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17 | 1.0 | 沈阳阴茎里面有疙瘩 - ```
沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17 | defect | 沈阳阴茎里面有疙瘩 沈阳阴茎里面有疙瘩〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at | 1 |
134,612 | 12,623,614,780 | IssuesEvent | 2020-06-14 00:10:01 | CoffeePerry/mercury-py | https://api.github.com/repos/CoffeePerry/mercury-py | opened | Load first documentation | documentation | # Load first documentation
Load to [README.md](https://github.com/CoffeePerry/mercury-py/blob/master/README.md):
- Prerequisites;
- First configurations;
- Future;
- More info.
Load all to [wiki](https://github.com/CoffeePerry/mercury-py/wiki). | 1.0 | Load first documentation - # Load first documentation
Load to [README.md](https://github.com/CoffeePerry/mercury-py/blob/master/README.md):
- Prerequisites;
- First configurations;
- Future;
- More info.
Load all to [wiki](https://github.com/CoffeePerry/mercury-py/wiki). | non_defect | load first documentation load first documentation load to prerequisites first configurations future more info load all to | 0 |
257,565 | 19,523,916,316 | IssuesEvent | 2021-12-30 01:53:41 | BlazorComponent/MASA.Blazor | https://api.github.com/repos/BlazorComponent/MASA.Blazor | closed | 【UI组件】Text fields - 示例,清空/修改文本框的内容后,点击空白区域,文本框内容被重置 | bug documentation resolved | 隐藏详细信息 - Another input ,输入文本后,点击空白区域,输入的内容被清空

| 1.0 | 【UI组件】Text fields - 示例,清空/修改文本框的内容后,点击空白区域,文本框内容被重置 - 隐藏详细信息 - Another input ,输入文本后,点击空白区域,输入的内容被清空

| non_defect | 【ui组件】text fields 示例,清空 修改文本框的内容后,点击空白区域,文本框内容被重置 隐藏详细信息 another input 输入文本后,点击空白区域,输入的内容被清空 | 0 |
776,686 | 27,264,560,939 | IssuesEvent | 2023-02-22 17:03:11 | ascheid/itsg33-pbmm-issue-gen | https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen | opened | CP-2(8): Contingency Plan | Identify Critical Assets | Priority: P3 Suggested Assignment: IT Security Function ITSG-33 Class: Operational Control: CP-2 | # Control Definition
CONTINGENCY PLAN | IDENTIFY CRITICAL ASSETS
The organization identifies critical information system assets supporting essential missions and business functions.
# Class
Operational
# Supplemental Guidance
Organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including, for example, as part of business impact analyses. Organizations identify critical information system assets so that additional safeguards and countermeasures can be employed (above and beyond those safeguards and countermeasures routinely implemented) to help ensure that organizational missions/business functions can continue to be conducted during contingency operations. In addition, the identification of critical information assets facilitates the prioritization of organizational resources. Critical information system assets include technical and operational aspects. Technical aspects include, for example, information technology services, information system components, information technology products, and mechanisms. Operational aspects include, for example, procedures (manually executed operations) and personnel (individuals operating technical safeguards and/or executing manual procedures). Organizational program protection plans can provide assistance in identifying critical assets. Related controls: SA-14, SA-15.
# Suggested Assignment
IT Security Function
# Support Teams
IT Operations Group
| 1.0 | CP-2(8): Contingency Plan | Identify Critical Assets - # Control Definition
CONTINGENCY PLAN | IDENTIFY CRITICAL ASSETS
The organization identifies critical information system assets supporting essential missions and business functions.
# Class
Operational
# Supplemental Guidance
Organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including, for example, as part of business impact analyses. Organizations identify critical information system assets so that additional safeguards and countermeasures can be employed (above and beyond those safeguards and countermeasures routinely implemented) to help ensure that organizational missions/business functions can continue to be conducted during contingency operations. In addition, the identification of critical information assets facilitates the prioritization of organizational resources. Critical information system assets include technical and operational aspects. Technical aspects include, for example, information technology services, information system components, information technology products, and mechanisms. Operational aspects include, for example, procedures (manually executed operations) and personnel (individuals operating technical safeguards and/or executing manual procedures). Organizational program protection plans can provide assistance in identifying critical assets. Related controls: SA-14, SA-15.
# Suggested Assignment
IT Security Function
# Support Teams
IT Operations Group
| non_defect | cp contingency plan identify critical assets control definition contingency plan identify critical assets the organization identifies critical information system assets supporting essential missions and business functions class operational supplemental guidance organizations may choose to carry out the contingency planning activities in this control enhancement as part of organizational business continuity planning including for example as part of business impact analyses organizations identify critical information system assets so that additional safeguards and countermeasures can be employed above and beyond those safeguards and countermeasures routinely implemented to help ensure that organizational missions business functions can continue to be conducted during contingency operations in addition the identification of critical information assets facilitates the prioritization of organizational resources critical information system assets include technical and operational aspects technical aspects include for example information technology services information system components information technology products and mechanisms operational aspects include for example procedures manually executed operations and personnel individuals operating technical safeguards and or executing manual procedures organizational program protection plans can provide assistance in identifying critical assets related controls sa sa suggested assignment it security function support teams it operations group | 0 |
307,159 | 26,518,545,168 | IssuesEvent | 2023-01-18 23:19:01 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | DISABLED test_graph_break_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) | module: flaky-tests skipped module: unknown | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_graph_break_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_graph_break_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301 | 1.0 | DISABLED test_graph_break_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_graph_break_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_graph_break_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301 | non_defect | disabled test graph break dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test graph break dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode | 0 |
73,767 | 7,354,580,655 | IssuesEvent | 2018-03-09 07:38:10 | FullScreenShenanigans/FullScreenPokemon | https://api.github.com/repos/FullScreenShenanigans/FullScreenPokemon | opened | Unit test gaining experience in battle | Good First Issue Help Wanted Test Coverage | When an enemy Pokemon is defeated, it should give experience to the Pokemon that defeated it. Some suggestions of cases to test (or file separate issues for if they're not yet implemented):
* One of your Pokemon defeated it
* >=2 of your Pokemon defeated it
* Exp share
* Exp all | 1.0 | Unit test gaining experience in battle - When an enemy Pokemon is defeated, it should give experience to the Pokemon that defeated it. Some suggestions of cases to test (or file separate issues for if they're not yet implemented):
* One of your Pokemon defeated it
* >=2 of your Pokemon defeated it
* Exp share
* Exp all | non_defect | unit test gaining experience in battle when an enemy pokemon is defeated it should give experience to the pokemon that defeated it some suggestions of cases to test or file separate issues for if they re not yet implemented one of your pokemon defeated it of your pokemon defeated it exp share exp all | 0 |
33,662 | 2,770,762,849 | IssuesEvent | 2015-05-01 16:53:05 | Metaswitch/gemini | https://api.github.com/repos/Metaswitch/gemini | closed | g.3gpp.ics media feature tag rendering is wrong | bug cat:easy medium-priority | When used in Accept-Contact/Contact/Reject-Contact headers, media tags are rendered with a leading ‘+’:
Accept-Contact: *;+g.3gpp.ics
Gemini does not include the ‘+’, it should. | 1.0 | g.3gpp.ics media feature tag rendering is wrong - When used in Accept-Contact/Contact/Reject-Contact headers, media tags are rendered with a leading ‘+’:
Accept-Contact: *;+g.3gpp.ics
Gemini does not include the ‘+’, it should. | non_defect | g ics media feature tag rendering is wrong when used in accept contact contact reject contact headers media tags are rendered with a leading ‘ ’ accept contact g ics gemini does not include the ‘ ’ it should | 0 |
22,177 | 3,609,714,283 | IssuesEvent | 2016-02-05 00:11:45 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | dart2js_extra/deferred_custom_loader_test fails in csp mode on browsers | area-dart2js Type-Defect | The test dart2js_extra/deferred_custom_loader_test, located at tests/compiler/dart2js_extra/deferred_custom_loader_test.dart, is failing
when run on linux content_shell in CSP mode, proabably for obvious reasons.
If this is expected, then the status should be changed to SkipByDesign, as it is for the deferred_fail_and_retry_test.
I'm doing that now, marking it SkipByDesign, with this issue number cited.
So if this is as expected, and not going to change, just close the issue and leave the status
unchanged.
Here is the error:
FAILED: dart2js-drt-checked release_ia32 dart2js_extra/deferred_custom_loader_test
Expected: Pass
Actual: RuntimeError
CommandOutput[dart2js]:
CONSOLE ERROR: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: "script-src 'self'".
CONSOLE MESSAGE: line 78: {
"type": "window_onerror",
"value": "window.onerror called: \n\nhttp://127.0.0.1:52467/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_deferred_custom_loader_test/test.js:5045:\nUncaught EvalError: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: \"script-src 'self'\".\n\n\n",
"timestamp": 0.087
Short reproduction command (experimental):
python tools/test.py -mrelease -cdart2js -rdrt -aia32 --use-sdk --write-debug-log --write-test-outcome-log --clear_browser_cache --minified --dart2js-batch --csp --checked -t480 dart2js_extra/deferred_custom_loader_test
| 1.0 | dart2js_extra/deferred_custom_loader_test fails in csp mode on browsers - The test dart2js_extra/deferred_custom_loader_test, located at tests/compiler/dart2js_extra/deferred_custom_loader_test.dart, is failing
when run on linux content_shell in CSP mode, proabably for obvious reasons.
If this is expected, then the status should be changed to SkipByDesign, as it is for the deferred_fail_and_retry_test.
I'm doing that now, marking it SkipByDesign, with this issue number cited.
So if this is as expected, and not going to change, just close the issue and leave the status
unchanged.
Here is the error:
FAILED: dart2js-drt-checked release_ia32 dart2js_extra/deferred_custom_loader_test
Expected: Pass
Actual: RuntimeError
CommandOutput[dart2js]:
CONSOLE ERROR: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: "script-src 'self'".
CONSOLE MESSAGE: line 78: {
"type": "window_onerror",
"value": "window.onerror called: \n\nhttp://127.0.0.1:52467/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_deferred_custom_loader_test/test.js:5045:\nUncaught EvalError: Refused to evaluate a string as JavaScript because 'unsafe-eval' is not an allowed source of script in the following Content Security Policy directive: \"script-src 'self'\".\n\n\n",
"timestamp": 0.087
Short reproduction command (experimental):
python tools/test.py -mrelease -cdart2js -rdrt -aia32 --use-sdk --write-debug-log --write-test-outcome-log --clear_browser_cache --minified --dart2js-batch --csp --checked -t480 dart2js_extra/deferred_custom_loader_test
| defect | extra deferred custom loader test fails in csp mode on browsers the test extra deferred custom loader test located at tests compiler extra deferred custom loader test dart is failing when run on linux content shell in csp mode proabably for obvious reasons if this is expected then the status should be changed to skipbydesign as it is for the deferred fail and retry test i m doing that now marking it skipbydesign with this issue number cited so if this is as expected and not going to change just close the issue and leave the status unchanged here is the error failed drt checked release extra deferred custom loader test expected pass actual runtimeerror commandoutput console error refused to evaluate a string as javascript because unsafe eval is not an allowed source of script in the following content security policy directive script src self console message line type window onerror value window onerror called n n evalerror refused to evaluate a string as javascript because unsafe eval is not an allowed source of script in the following content security policy directive script src self n n n timestamp short reproduction command experimental python tools test py mrelease rdrt use sdk write debug log write test outcome log clear browser cache minified batch csp checked extra deferred custom loader test | 1 |
21,174 | 3,466,971,940 | IssuesEvent | 2015-12-22 08:41:03 | Ryzhehvost/keyla | https://api.github.com/repos/Ryzhehvost/keyla | closed | keyla-0.1.9-x64 would not start under win7-64 | auto-migrated duplicate Priority-Medium Type-Defect | ```
1. Install keyla-0.1.9-x64-setup.exe
2. Run keyla.exe
Application immediately crushes with 0xc000007b
keyla-0.1.9-x64, Win7 Pro 64 bit
```
Original issue reported on code.google.com by `smishc...@gmail.com` on 9 Jan 2013 at 2:24
Attachments:
* [1.PNG](https://storage.googleapis.com/google-code-attachments/keyla/issue-28/comment-0/1.PNG)
| 1.0 | keyla-0.1.9-x64 would not start under win7-64 - ```
1. Install keyla-0.1.9-x64-setup.exe
2. Run keyla.exe
Application immediately crushes with 0xc000007b
keyla-0.1.9-x64, Win7 Pro 64 bit
```
Original issue reported on code.google.com by `smishc...@gmail.com` on 9 Jan 2013 at 2:24
Attachments:
* [1.PNG](https://storage.googleapis.com/google-code-attachments/keyla/issue-28/comment-0/1.PNG)
| defect | keyla would not start under install keyla setup exe run keyla exe application immediately crushes with keyla pro bit original issue reported on code google com by smishc gmail com on jan at attachments | 1 |
22,864 | 3,727,389,257 | IssuesEvent | 2016-03-06 08:04:54 | godfather1103/mentohust | https://api.github.com/repos/godfather1103/mentohust | closed | mipsel版mentohust认证锐捷4.85,提示“用户不允许使用本服务” | auto-migrated Priority-Medium Type-Defect | ```
使用Windows版Mentohust认证一切正常,但是使用Mipsel版用相同的�
��数和方式进行认证,会收到“用户不允许使用本服务”的提
示。
返回结果如下
** Netcard: vlan2
** Auth timeout(s): 8 s
** Heartbeat interval(s): 30 s
** Failure waittime(s): 15 s
** Allow failures: 8 times
** Multicast addr: Ruijie
** DHCP Type: Disabled
** MAC: 94:44:52:af:e7:64
** IP: 0.0.0.0
** Netmask: 0.0.0.0
!! Save the authentication parameters to /jffs/etc/mentohust.conf failure锛?
!! Save the authentication parameters to /tmp/root/mentohust.conf failure锛?
>> Looking for server...
** Authentication MAC: 00:1a:a9:1c:9b:c6
>> Send user name...
>> Send password...
>> failure!
$$ System hint: 用户不允许使用本服务!
最后一次正常使用是在两个月前,锐捷的版本为4.85,并未更�
��
```
Original issue reported on code.google.com by `Xyzh...@gmail.com` on 31 Aug 2013 at 4:30 | 1.0 | mipsel版mentohust认证锐捷4.85,提示“用户不允许使用本服务” - ```
使用Windows版Mentohust认证一切正常,但是使用Mipsel版用相同的�
��数和方式进行认证,会收到“用户不允许使用本服务”的提
示。
返回结果如下
** Netcard: vlan2
** Auth timeout(s): 8 s
** Heartbeat interval(s): 30 s
** Failure waittime(s): 15 s
** Allow failures: 8 times
** Multicast addr: Ruijie
** DHCP Type: Disabled
** MAC: 94:44:52:af:e7:64
** IP: 0.0.0.0
** Netmask: 0.0.0.0
!! Save the authentication parameters to /jffs/etc/mentohust.conf failure锛?
!! Save the authentication parameters to /tmp/root/mentohust.conf failure锛?
>> Looking for server...
** Authentication MAC: 00:1a:a9:1c:9b:c6
>> Send user name...
>> Send password...
>> failure!
$$ System hint: 用户不允许使用本服务!
最后一次正常使用是在两个月前,锐捷的版本为4.85,并未更�
��
```
Original issue reported on code.google.com by `Xyzh...@gmail.com` on 31 Aug 2013 at 4:30 | defect | ,提示“用户不允许使用本服务” 使用windows版mentohust认证一切正常,但是使用mipsel版用相同的� ��数和方式进行认证,会收到“用户不允许使用本服务”的提 示。 返回结果如下 netcard auth timeout s s heartbeat interval s s failure waittime s s allow failures times multicast addr ruijie dhcp type disabled mac af ip netmask save the authentication parameters to jffs etc mentohust conf failure锛 save the authentication parameters to tmp root mentohust conf failure锛 looking for server authentication mac send user name send password failure system hint 用户不允许使用本服务 最后一次正常使用是在两个月前, ,并未更� �� original issue reported on code google com by xyzh gmail com on aug at | 1 |
59,059 | 17,015,345,758 | IssuesEvent | 2021-07-02 11:11:28 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | opened | When viewing "My Tracks" up and down arrows affect the window behind | Component: potlatch2 Priority: minor Type: defect | **[Submitted to the original trac issue database at 2.01pm, Saturday, 22nd January 2011]**
To reproduce:
Open Potlatch2
Select "My Tracks" from the top left
Attempt to scroll down (or up) the list with the arrow keys
Nothing obvious seems to happen, but in fact the (smudged-out) map behind is moving. When the "My Tracks" box is closed, you're somewhere that you don't expect to be.
If the arrow keys are supposed to move the map behind then it presumably shouldn't be smudged-out? Also some kind of clue about what the arrow keys do here would be useful. If they're not, how about making them just, er, scroll up and down the list in the foreground?
| 1.0 | When viewing "My Tracks" up and down arrows affect the window behind - **[Submitted to the original trac issue database at 2.01pm, Saturday, 22nd January 2011]**
To reproduce:
Open Potlatch2
Select "My Tracks" from the top left
Attempt to scroll down (or up) the list with the arrow keys
Nothing obvious seems to happen, but in fact the (smudged-out) map behind is moving. When the "My Tracks" box is closed, you're somewhere that you don't expect to be.
If the arrow keys are supposed to move the map behind then it presumably shouldn't be smudged-out? Also some kind of clue about what the arrow keys do here would be useful. If they're not, how about making them just, er, scroll up and down the list in the foreground?
| defect | when viewing my tracks up and down arrows affect the window behind to reproduce open select my tracks from the top left attempt to scroll down or up the list with the arrow keys nothing obvious seems to happen but in fact the smudged out map behind is moving when the my tracks box is closed you re somewhere that you don t expect to be if the arrow keys are supposed to move the map behind then it presumably shouldn t be smudged out also some kind of clue about what the arrow keys do here would be useful if they re not how about making them just er scroll up and down the list in the foreground | 1 |
59,107 | 17,015,744,334 | IssuesEvent | 2021-07-02 11:46:58 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | opened | proposed route=bicycle relation overrides normal cycle route | Component: opencyclemap Priority: major Type: defect | **[Submitted to the original trac issue database at 9.28pm, Saturday, 5th November 2011]**
http://www.openstreetmap.org/?lat=28.513&lon=-81.4106&zoom=13&layers=C
http://www.openstreetmap.org/browse/way/11182207
The street south of the lake in the center is tagged lcn=yes and is part of an lcn=proposed relation. It should be displayed as normal lcn, not proposed. | 1.0 | proposed route=bicycle relation overrides normal cycle route - **[Submitted to the original trac issue database at 9.28pm, Saturday, 5th November 2011]**
http://www.openstreetmap.org/?lat=28.513&lon=-81.4106&zoom=13&layers=C
http://www.openstreetmap.org/browse/way/11182207
The street south of the lake in the center is tagged lcn=yes and is part of an lcn=proposed relation. It should be displayed as normal lcn, not proposed. | defect | proposed route bicycle relation overrides normal cycle route the street south of the lake in the center is tagged lcn yes and is part of an lcn proposed relation it should be displayed as normal lcn not proposed | 1 |
29,279 | 8,316,443,215 | IssuesEvent | 2018-09-25 09:03:52 | Microsoft/vscode | https://api.github.com/repos/Microsoft/vscode | closed | Search breaks when keybindings editor is open | broken-build important search | - Open keybindings editor
- Search in search viewlet
- Spins forever
```
TypeError: Cannot read property 'toString' of null
errors.ts:51
at EditorService.doGetOpened (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:374:38)
at EditorService.isOpen (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:335:27)
at file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:298:46
at Array.forEach (<anonymous>)
at SearchService.getLocalResults (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:293:24)
at SearchService.search (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:97:37)
at SearchModel.search (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/common/searchModel.js:743:53)
at SearchView.onQueryTriggered (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:1095:28)
at file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:886:23
```
Looks like `resource` can be null here https://github.com/Microsoft/vscode/blob/master/src/vs/workbench/services/editor/browser/editorService.ts#L422
This breaks smoketests too https://dev.azure.com/vscode/VSCode/_build/results?buildId=7711&view=logs | 1.0 | Search breaks when keybindings editor is open - - Open keybindings editor
- Search in search viewlet
- Spins forever
```
TypeError: Cannot read property 'toString' of null
errors.ts:51
at EditorService.doGetOpened (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:374:38)
at EditorService.isOpen (file:///Users/roblou/code/vscode/out/vs/workbench/services/editor/browser/editorService.js:335:27)
at file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:298:46
at Array.forEach (<anonymous>)
at SearchService.getLocalResults (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:293:24)
at SearchService.search (file:///Users/roblou/code/vscode/out/vs/workbench/services/search/node/searchService.js:97:37)
at SearchModel.search (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/common/searchModel.js:743:53)
at SearchView.onQueryTriggered (file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:1095:28)
at file:///Users/roblou/code/vscode/out/vs/workbench/parts/search/browser/searchView.js:886:23
```
Looks like `resource` can be null here https://github.com/Microsoft/vscode/blob/master/src/vs/workbench/services/editor/browser/editorService.ts#L422
This breaks smoketests too https://dev.azure.com/vscode/VSCode/_build/results?buildId=7711&view=logs | non_defect | search breaks when keybindings editor is open open keybindings editor search in search viewlet spins forever typeerror cannot read property tostring of null errors ts at editorservice dogetopened file users roblou code vscode out vs workbench services editor browser editorservice js at editorservice isopen file users roblou code vscode out vs workbench services editor browser editorservice js at file users roblou code vscode out vs workbench services search node searchservice js at array foreach at searchservice getlocalresults file users roblou code vscode out vs workbench services search node searchservice js at searchservice search file users roblou code vscode out vs workbench services search node searchservice js at searchmodel search file users roblou code vscode out vs workbench parts search common searchmodel js at searchview onquerytriggered file users roblou code vscode out vs workbench parts search browser searchview js at file users roblou code vscode out vs workbench parts search browser searchview js looks like resource can be null here this breaks smoketests too | 0 |
164,762 | 12,812,914,181 | IssuesEvent | 2020-07-04 09:36:53 | aliasrobotics/RVD | https://api.github.com/repos/aliasrobotics/RVD | closed | RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80 | CWE-134 bug components software flawfinder flawfinder_level_4 mitigated robot component: PX4 static analysis testing triage version: v1.8.0 version: v1.9.0 | ```yaml
id: 2720
title: 'RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker,
they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80'
type: bug
description: If format strings can be influenced by an attacker, they can be exploited
(CWE-134). Use a constant for the format specification. . Happening @ ...4_layer/px4_qurt_impl.cpp:80
cwe:
- CWE-134
cve: None
keywords:
- flawfinder
- flawfinder_level_4
- static analysis
- testing
- triage
- CWE-134
- bug
- 'version: v1.9.0'
- 'robot component: PX4'
- components software
system: ./Firmware/platforms/qurt/src/px4_layer/px4_qurt_impl.cpp:80:2
vendor: null
severity:
rvss-score: 0
rvss-vector: ''
severity-description: ''
cvss-score: 0
cvss-vector: ''
links:
- https://github.com/aliasrobotics/RVD/issues/2720
flaw:
phase: testing
specificity: subject-specific
architectural-location: application-specific
application: N/A
subsystem: N/A
package: N/A
languages: None
date-detected: 2020-06-29 (15:05)
detected-by: Alias Robotics
detected-by-method: testing static
date-reported: 2020-06-29 (15:05)
reported-by: Alias Robotics
reported-by-relationship: automatic
issue: https://github.com/aliasrobotics/RVD/issues/2720
reproducibility: always
trace: (context) \tprintf(fmt, args);
reproduction: See artifacts below (if available)
reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download
exploitation:
description: ''
exploitation-image: ''
exploitation-vector: ''
exploitation-recipe: ''
mitigation:
description: Use a constant for the format specification
pull-request: ''
date-mitigation: ''
``` | 1.0 | RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker, they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80 - ```yaml
id: 2720
title: 'RVD#2720: CWE-134 (format), If format strings can be influenced by an attacker,
they can be exploi... @ 4_layer/px4_qurt_impl.cpp:80'
type: bug
description: If format strings can be influenced by an attacker, they can be exploited
(CWE-134). Use a constant for the format specification. . Happening @ ...4_layer/px4_qurt_impl.cpp:80
cwe:
- CWE-134
cve: None
keywords:
- flawfinder
- flawfinder_level_4
- static analysis
- testing
- triage
- CWE-134
- bug
- 'version: v1.9.0'
- 'robot component: PX4'
- components software
system: ./Firmware/platforms/qurt/src/px4_layer/px4_qurt_impl.cpp:80:2
vendor: null
severity:
rvss-score: 0
rvss-vector: ''
severity-description: ''
cvss-score: 0
cvss-vector: ''
links:
- https://github.com/aliasrobotics/RVD/issues/2720
flaw:
phase: testing
specificity: subject-specific
architectural-location: application-specific
application: N/A
subsystem: N/A
package: N/A
languages: None
date-detected: 2020-06-29 (15:05)
detected-by: Alias Robotics
detected-by-method: testing static
date-reported: 2020-06-29 (15:05)
reported-by: Alias Robotics
reported-by-relationship: automatic
issue: https://github.com/aliasrobotics/RVD/issues/2720
reproducibility: always
trace: (context) \tprintf(fmt, args);
reproduction: See artifacts below (if available)
reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download
exploitation:
description: ''
exploitation-image: ''
exploitation-vector: ''
exploitation-recipe: ''
mitigation:
description: Use a constant for the format specification
pull-request: ''
date-mitigation: ''
``` | non_defect | rvd cwe format if format strings can be influenced by an attacker they can be exploi layer qurt impl cpp yaml id title rvd cwe format if format strings can be influenced by an attacker they can be exploi layer qurt impl cpp type bug description if format strings can be influenced by an attacker they can be exploited cwe use a constant for the format specification happening layer qurt impl cpp cwe cwe cve none keywords flawfinder flawfinder level static analysis testing triage cwe bug version robot component components software system firmware platforms qurt src layer qurt impl cpp vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace context tprintf fmt args reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline jobs artifacts download exploitation description exploitation image exploitation vector exploitation recipe mitigation description use a constant for the format specification pull request date mitigation | 0 |
65,756 | 19,680,702,156 | IssuesEvent | 2022-01-11 16:29:08 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | opened | taking a video call does not always default to loudspeaker | T-Defect | ### Steps to reproduce
call somebody on their iphone.
they're not so technologically inclined as to open this report themselves. i'll see if i can get version numbers, but it should be decently updated as iOS does.
### Outcome
#### What did you expect?
default to loudspeaker
#### What happened instead?
earphone speaker is selected by default
### Your phone model
couple years old
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | taking a video call does not always default to loudspeaker - ### Steps to reproduce
call somebody on their iphone.
they're not so technologically inclined as to open this report themselves. i'll see if i can get version numbers, but it should be decently updated as iOS does.
### Outcome
#### What did you expect?
default to loudspeaker
#### What happened instead?
earphone speaker is selected by default
### Your phone model
couple years old
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | defect | taking a video call does not always default to loudspeaker steps to reproduce call somebody on their iphone they re not so technologically inclined as to open this report themselves i ll see if i can get version numbers but it should be decently updated as ios does outcome what did you expect default to loudspeaker what happened instead earphone speaker is selected by default your phone model couple years old operating system version no response application version no response homeserver no response will you send logs no | 1 |
1,708 | 2,603,969,851 | IssuesEvent | 2015-02-24 18:59:58 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳龟头有豆豆怎么回事 | auto-migrated Priority-Medium Type-Defect | ```
沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:25 | 1.0 | 沈阳龟头有豆豆怎么回事 - ```
沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:25 | defect | 沈阳龟头有豆豆怎么回事 沈阳龟头有豆豆怎么回事〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at | 1 |
78,269 | 27,403,861,125 | IssuesEvent | 2023-03-01 04:17:55 | pageseeder/diffx | https://api.github.com/repos/pageseeder/diffx | closed | Please add to Maven Central | Priority-Medium Type-Defect auto-migrated | ```
Please add wo-diffx to Maven Central.
You can follow this guide:
https://docs.sonatype.org/display/Repository/Sonatype+OSS+Maven+Repository+Usage
+Guide
thanks .. Jason
```
Original issue reported on code.google.com by `jhar...@gmail.com` on 27 May 2013 at 1:35
| 1.0 | Please add to Maven Central - ```
Please add wo-diffx to Maven Central.
You can follow this guide:
https://docs.sonatype.org/display/Repository/Sonatype+OSS+Maven+Repository+Usage
+Guide
thanks .. Jason
```
Original issue reported on code.google.com by `jhar...@gmail.com` on 27 May 2013 at 1:35
| defect | please add to maven central please add wo diffx to maven central you can follow this guide guide thanks jason original issue reported on code google com by jhar gmail com on may at | 1 |
21,862 | 3,573,522,110 | IssuesEvent | 2016-01-27 07:02:03 | hunnee/jsontoken | https://api.github.com/repos/hunnee/jsontoken | closed | UrlBasedVerifierProvider does not work with recent version of gson | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1.
In pom.xml switch to a recent version of gson, 2.2.4 for example
2.
Use UrlBasedVerifierProvider to fetch a certicate from an url.
3.
What is the expected output? What do you see instead?
I should get a List<Verifier>
I get an exception
[INFO] com.google.gson.JsonSyntaxException:
com.google.gson.stream.MalformedJsonException: Use JsonReader.setLenient(true)
to accept malformed JSON at line 2 column 2
[INFO] at com.google.gson.JsonParser.parse(JsonParser.java:65)
[INFO] at com.google.gson.JsonParser.parse(JsonParser.java:45)
[INFO] at
net.oauth.jsontoken.discovery.UrlBasedVerifierProvider.findVerifier(UrlBasedVeri
fierProvider.java:58)
[INFO] at
net.oauth.jsontoken.JsonTokenParser.verifyAndDeserialize(JsonTokenParser.java:10
8)
How to Fix it
@
https://code.google.com/p/jsontoken/source/browse/trunk/src/main/java/net/oauth/
jsontoken/discovery/UrlBasedVerifierProvider.java#51
replace
String line = "";
do {
line = buff.readLine();
content.append(line + "\n");
} while (line != null);
with
String line = buff.readLine();
while (line != null) {
content.append(line + "\n");
line = buff.readLine();
} ;
The later won't add an unwanted "null" string at the end of the json string.
```
Original issue reported on code.google.com by `aurelien...@gmail.com` on 22 Aug 2013 at 8:10 | 1.0 | UrlBasedVerifierProvider does not work with recent version of gson - ```
What steps will reproduce the problem?
1.
In pom.xml switch to a recent version of gson, 2.2.4 for example
2.
Use UrlBasedVerifierProvider to fetch a certicate from an url.
3.
What is the expected output? What do you see instead?
I should get a List<Verifier>
I get an exception
[INFO] com.google.gson.JsonSyntaxException:
com.google.gson.stream.MalformedJsonException: Use JsonReader.setLenient(true)
to accept malformed JSON at line 2 column 2
[INFO] at com.google.gson.JsonParser.parse(JsonParser.java:65)
[INFO] at com.google.gson.JsonParser.parse(JsonParser.java:45)
[INFO] at
net.oauth.jsontoken.discovery.UrlBasedVerifierProvider.findVerifier(UrlBasedVeri
fierProvider.java:58)
[INFO] at
net.oauth.jsontoken.JsonTokenParser.verifyAndDeserialize(JsonTokenParser.java:10
8)
How to Fix it
@
https://code.google.com/p/jsontoken/source/browse/trunk/src/main/java/net/oauth/
jsontoken/discovery/UrlBasedVerifierProvider.java#51
replace
String line = "";
do {
line = buff.readLine();
content.append(line + "\n");
} while (line != null);
with
String line = buff.readLine();
while (line != null) {
content.append(line + "\n");
line = buff.readLine();
} ;
The later won't add an unwanted "null" string at the end of the json string.
```
Original issue reported on code.google.com by `aurelien...@gmail.com` on 22 Aug 2013 at 8:10 | defect | urlbasedverifierprovider does not work with recent version of gson what steps will reproduce the problem in pom xml switch to a recent version of gson for example use urlbasedverifierprovider to fetch a certicate from an url what is the expected output what do you see instead i should get a list i get an exception com google gson jsonsyntaxexception com google gson stream malformedjsonexception use jsonreader setlenient true to accept malformed json at line column at com google gson jsonparser parse jsonparser java at com google gson jsonparser parse jsonparser java at net oauth jsontoken discovery urlbasedverifierprovider findverifier urlbasedveri fierprovider java at net oauth jsontoken jsontokenparser verifyanddeserialize jsontokenparser java how to fix it jsontoken discovery urlbasedverifierprovider java replace string line do line buff readline content append line n while line null with string line buff readline while line null content append line n line buff readline the later won t add an unwanted null string at the end of the json string original issue reported on code google com by aurelien gmail com on aug at | 1 |
2,434 | 11,950,542,903 | IssuesEvent | 2020-04-03 15:22:18 | gcdevops/HRWhiteListing | https://api.github.com/repos/gcdevops/HRWhiteListing | closed | Automate creation of custom fields and subfields | automation | Creation of fields needs to be automated
Related to #76 | 1.0 | Automate creation of custom fields and subfields - Creation of fields needs to be automated
Related to #76 | non_defect | automate creation of custom fields and subfields creation of fields needs to be automated related to | 0 |
49,672 | 13,187,249,365 | IssuesEvent | 2020-08-13 02:49:12 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | PolyplopiaSegment issue at Detector Level (Trac #1851) | Incomplete Migration Migrated from Trac combo simulation defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1851">https://code.icecube.wisc.edu/ticket/1851</a>, reported by saxani and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:54",
"description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1550067174476394",
"component": "combo simulation",
"summary": "PolyplopiaSegment issue at Detector Level",
"priority": "normal",
"keywords": "",
"time": "2016-09-06T14:31:56",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
| 1.0 | PolyplopiaSegment issue at Detector Level (Trac #1851) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1851">https://code.icecube.wisc.edu/ticket/1851</a>, reported by saxani and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:54",
"description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1550067174476394",
"component": "combo simulation",
"summary": "PolyplopiaSegment issue at Detector Level",
"priority": "normal",
"keywords": "",
"time": "2016-09-06T14:31:56",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
| defect | polyplopiasegment issue at detector level trac migrated from json status closed changetime description hi n ni m trying to run polyplopiasegment on the simulation release at detector level i end up with this error n n data user saxani environments builddet env shell sh line bus error path path ld library path ld library path dyld library path dyld library path pythonpath pythonpath icetray classpath icetray classpath platform src src build build ports ports testdata testdata rootsys cvmfs icecube opensciencegrid org rhel root shell shell new shell argv n ncan we reimplement the older version of polyplopiasegment into this release so that it can run after photon propagation n nattn juan carlos n nthanks n nspencer reporter saxani cc resolution fixed ts component combo simulation summary polyplopiasegment issue at detector level priority normal keywords time milestone owner juancarlos type defect | 1 |
25,056 | 4,186,289,305 | IssuesEvent | 2016-06-23 14:06:07 | jbs1/testingtrac | https://api.github.com/repos/jbs1/testingtrac | closed | tabbing and friends (Trac #1369) | bindings defect Incomplete Migration Migrated from Trac | Migrated from https://trac.kwarc.info/ticket/1369
```json
{
"changetime": "2011-07-11T13:45:23",
"cc": "miller",
"blockedby": "",
"keywords": "",
"owner": "stamer",
"due_close": "",
"severity": "normal",
"include_gantt": "0",
"priority": "high",
"due_assign": "",
"type": "defect",
"status": "closed",
"description": "in math/papers/9901119, we find \\begin{tabbing} and\\kill, they do not work yet. \n\nI suspect they are TeX builtins",
"reporter": "kohlhase",
"_ts": "1310391923000000",
"component": "bindings",
"dependencies": "",
"milestone": "",
"blocking": "",
"summary": "tabbing and friends",
"time": "2007-01-22T17:16:45",
"resolution": "fixed"
}
```
| 1.0 | tabbing and friends (Trac #1369) - Migrated from https://trac.kwarc.info/ticket/1369
```json
{
"changetime": "2011-07-11T13:45:23",
"cc": "miller",
"blockedby": "",
"keywords": "",
"owner": "stamer",
"due_close": "",
"severity": "normal",
"include_gantt": "0",
"priority": "high",
"due_assign": "",
"type": "defect",
"status": "closed",
"description": "in math/papers/9901119, we find \\begin{tabbing} and\\kill, they do not work yet. \n\nI suspect they are TeX builtins",
"reporter": "kohlhase",
"_ts": "1310391923000000",
"component": "bindings",
"dependencies": "",
"milestone": "",
"blocking": "",
"summary": "tabbing and friends",
"time": "2007-01-22T17:16:45",
"resolution": "fixed"
}
```
| defect | tabbing and friends trac migrated from json changetime cc miller blockedby keywords owner stamer due close severity normal include gantt priority high due assign type defect status closed description in math papers we find begin tabbing and kill they do not work yet n ni suspect they are tex builtins reporter kohlhase ts component bindings dependencies milestone blocking summary tabbing and friends time resolution fixed | 1 |
1,130 | 2,596,584,097 | IssuesEvent | 2015-02-20 21:44:06 | STEllAR-GROUP/hpx | https://api.github.com/repos/STEllAR-GROUP/hpx | closed | Compilation error with papi | category: components type: defect | ```
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: error: ‘function_nonser’ is not a member of ‘hpx::performance_counters::papi::util’
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: note: suggested alternative:
In file included from /home/pagrubel/hpx/hpx/util/function.hpp:13:0,
from /home/pagrubel/hpx/hpx/hpx_fwd.hpp:43,
from /home/pagrubel/hpx/hpx/hpx.hpp:9,
from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15:
/home/pagrubel/hpx/hpx/util/detail/function_template.hpp:343:54: note: ‘hpx::util::function_nonser’
using function_nonser = function<Sig, void, void>;
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:55: error: ‘startup_func’ was not declared in this scope
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
In file included from /usr/include/c++/4.9/bits/atomic_base.h:36:0,
from /usr/include/c++/4.9/atomic:41,
from /opt/boost/1.56.0-release/boost/lockfree/detail/atomic.hpp:32,
from /opt/boost/1.56.0-release/boost/lockfree/queue.hpp:19,
from /home/pagrubel/hpx/hpx/util/lockfree/freelist.hpp:15,
from /home/pagrubel/hpx/hpx/runtime/threads/thread_data.hpp:26,
from /home/pagrubel/hpx/hpx/include/threadmanager.hpp:15,
from /home/pagrubel/hpx/hpx/include/runtime.hpp:14,
from /home/pagrubel/hpx/hpx/hpx.hpp:11,
from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15:
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:69: error: expected primary-expression before ‘bool’
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:86: error: expression list treated as compound expression in initializer [-fpermissive]
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:289:5: error: expected ‘,’ or ‘;’ before ‘{’ token
{
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
hpx::performance_counters::papi::util::get_options_description);
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
/opt/boost/1.56.0-release/boost/format/feed_args.hpp: In instantiation of ‘void boost::io::detail::call_put_head(std::basic_ostream<_CharT, _Traits>&, const void*) [with Ch = char; Tr = std::char_traits<char>; T = long int]’:
/opt/boost/1.56.0-release/boost/format/feed_args.hpp:135:47: required from ‘boost::io::detail::put_holder<Ch, Tr>::put_holder(T&) [with T = long int; Ch = char; Tr = std::char_traits<char>]’
```
| 1.0 | Compilation error with papi - ```
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: error: ‘function_nonser’ is not a member of ‘hpx::performance_counters::papi::util’
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:24: note: suggested alternative:
In file included from /home/pagrubel/hpx/hpx/util/function.hpp:13:0,
from /home/pagrubel/hpx/hpx/hpx_fwd.hpp:43,
from /home/pagrubel/hpx/hpx/hpx.hpp:9,
from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15:
/home/pagrubel/hpx/hpx/util/detail/function_template.hpp:343:54: note: ‘hpx::util::function_nonser’
using function_nonser = function<Sig, void, void>;
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:55: error: ‘startup_func’ was not declared in this scope
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
In file included from /usr/include/c++/4.9/bits/atomic_base.h:36:0,
from /usr/include/c++/4.9/atomic:41,
from /opt/boost/1.56.0-release/boost/lockfree/detail/atomic.hpp:32,
from /opt/boost/1.56.0-release/boost/lockfree/queue.hpp:19,
from /home/pagrubel/hpx/hpx/util/lockfree/freelist.hpp:15,
from /home/pagrubel/hpx/hpx/runtime/threads/thread_data.hpp:26,
from /home/pagrubel/hpx/hpx/include/threadmanager.hpp:15,
from /home/pagrubel/hpx/hpx/include/runtime.hpp:14,
from /home/pagrubel/hpx/hpx/hpx.hpp:11,
from /home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:15:
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:69: error: expected primary-expression before ‘bool’
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:288:86: error: expression list treated as compound expression in initializer [-fpermissive]
bool check_startup(util::function_nonser<void()>& startup_func, bool& pre_startup)
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:289:5: error: expected ‘,’ or ‘;’ before ‘{’ token
{
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
hpx::performance_counters::papi::util::get_options_description);
^
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
/home/pagrubel/hpx/src/components/papi_counters/papi_startup.cpp:322:68: error: expected ‘}’ at end of input
/opt/boost/1.56.0-release/boost/format/feed_args.hpp: In instantiation of ‘void boost::io::detail::call_put_head(std::basic_ostream<_CharT, _Traits>&, const void*) [with Ch = char; Tr = std::char_traits<char>; T = long int]’:
/opt/boost/1.56.0-release/boost/format/feed_args.hpp:135:47: required from ‘boost::io::detail::put_holder<Ch, Tr>::put_holder(T&) [with T = long int; Ch = char; Tr = std::char_traits<char>]’
```
| defect | compilation error with papi home pagrubel hpx src components papi counters papi startup cpp error ‘function nonser’ is not a member of ‘hpx performance counters papi util’ bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp note suggested alternative in file included from home pagrubel hpx hpx util function hpp from home pagrubel hpx hpx hpx fwd hpp from home pagrubel hpx hpx hpx hpp from home pagrubel hpx src components papi counters papi startup cpp home pagrubel hpx hpx util detail function template hpp note ‘hpx util function nonser’ using function nonser function home pagrubel hpx src components papi counters papi startup cpp error ‘startup func’ was not declared in this scope bool check startup util function nonser startup func bool pre startup in file included from usr include c bits atomic base h from usr include c atomic from opt boost release boost lockfree detail atomic hpp from opt boost release boost lockfree queue hpp from home pagrubel hpx hpx util lockfree freelist hpp from home pagrubel hpx hpx runtime threads thread data hpp from home pagrubel hpx hpx include threadmanager hpp from home pagrubel hpx hpx include runtime hpp from home pagrubel hpx hpx hpx hpp from home pagrubel hpx src components papi counters papi startup cpp home pagrubel hpx src components papi counters papi startup cpp error expected primary expression before ‘bool’ bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp error expression list treated as compound expression in initializer bool check startup util function nonser startup func bool pre startup home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ or ‘ ’ before ‘ ’ token home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input hpx performance counters papi util get options description home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input home pagrubel hpx src components papi counters papi startup cpp error expected ‘ ’ at end of input opt boost release boost format feed args hpp in instantiation of ‘void boost io detail call put head std basic ostream const void ’ opt boost release boost format feed args hpp required from ‘boost io detail put holder put holder t ’ | 1 |
4,104 | 10,575,755,479 | IssuesEvent | 2019-10-07 16:21:22 | fac-17/Autsera | https://api.github.com/repos/fac-17/Autsera | closed | Router has one endpoint for multiple instances of a component, figure out how to pass props to router endpoint. | architecture | - [x] PlacePage component is rendered on the /page endpoint, but it needs to know which place to render.
- [x] it will then read values for that place from the global data array | 1.0 | Router has one endpoint for multiple instances of a component, figure out how to pass props to router endpoint. - - [x] PlacePage component is rendered on the /page endpoint, but it needs to know which place to render.
- [x] it will then read values for that place from the global data array | non_defect | router has one endpoint for multiple instances of a component figure out how to pass props to router endpoint placepage component is rendered on the page endpoint but it needs to know which place to render it will then read values for that place from the global data array | 0 |
77,069 | 26,752,554,746 | IssuesEvent | 2023-01-30 20:53:21 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | issue with keys | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see? Bisq General room
2. What do you click? Not clicking anything. Scanning messages in the room, I can see an apparent struggle to decrypt messages going on.
3. More steps…
My MacBook Pro started struggling to decrypt in an encrypted room and despite signing in and out of Element, I continue to receive the following error message "Some messages could not be decrypted
Unfortunately, there are no other verified devices to request decryption keys from. Signing in and verifying other devices may help avoid this situation in the future."
I left the room and tried to reenter and was urged to open an issue here due to an error.
### Outcome
#### What did you expect?
#### What happened instead?
error message keeps reappearing. Struggle to decrypt iMessages. Final error message recited above. Messages in room prior to January 24 no longer appear.
### Operating system
Mac OS 10.15.7
### Application version
Version 1.11.21-rc.1 (1.11.21-rc.1)
### How did you install the app?
App Store, probably. I downloaded last summer and had no issues until now
### Homeserver
Matrix
### Will you send logs?
Yes | 1.0 | issue with keys - ### Steps to reproduce
1. Where are you starting? What can you see? Bisq General room
2. What do you click? Not clicking anything. Scanning messages in the room, I can see an apparent struggle to decrypt messages going on.
3. More steps…
My MacBook Pro started struggling to decrypt in an encrypted room and despite signing in and out of Element, I continue to receive the following error message "Some messages could not be decrypted
Unfortunately, there are no other verified devices to request decryption keys from. Signing in and verifying other devices may help avoid this situation in the future."
I left the room and tried to reenter and was urged to open an issue here due to an error.
### Outcome
#### What did you expect?
#### What happened instead?
error message keeps reappearing. Struggle to decrypt iMessages. Final error message recited above. Messages in room prior to January 24 no longer appear.
### Operating system
Mac OS 10.15.7
### Application version
Version 1.11.21-rc.1 (1.11.21-rc.1)
### How did you install the app?
App Store, probably. I downloaded last summer and had no issues until now
### Homeserver
Matrix
### Will you send logs?
Yes | defect | issue with keys steps to reproduce where are you starting what can you see bisq general room what do you click not clicking anything scanning messages in the room i can see an apparent struggle to decrypt messages going on more steps… my macbook pro started struggling to decrypt in an encrypted room and despite signing in and out of element i continue to receive the following error message some messages could not be decrypted unfortunately there are no other verified devices to request decryption keys from signing in and verifying other devices may help avoid this situation in the future i left the room and tried to reenter and was urged to open an issue here due to an error outcome what did you expect what happened instead error message keeps reappearing struggle to decrypt imessages final error message recited above messages in room prior to january no longer appear operating system mac os application version version rc rc how did you install the app app store probably i downloaded last summer and had no issues until now homeserver matrix will you send logs yes | 1 |
543,434 | 15,881,942,773 | IssuesEvent | 2021-04-09 15:24:24 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Revision dropdown goes blank when creating a new version | API-M 4.0.0 Feature/Revisioning Priority/High React-UI Type/Bug | ### Description:
The revision select dropdown in the Top Menu bar in Publisher goes blank when creating a new API version. Current API option is not selected. It shows up after page reload.
<img width="1653" alt="Screenshot 2021-04-02 at 17 22 48" src="https://user-images.githubusercontent.com/8557410/113413587-a1b2ca00-93d8-11eb-80d1-55553102fe67.png">
| 1.0 | Revision dropdown goes blank when creating a new version - ### Description:
The revision select dropdown in the Top Menu bar in Publisher goes blank when creating a new API version. Current API option is not selected. It shows up after page reload.
<img width="1653" alt="Screenshot 2021-04-02 at 17 22 48" src="https://user-images.githubusercontent.com/8557410/113413587-a1b2ca00-93d8-11eb-80d1-55553102fe67.png">
| non_defect | revision dropdown goes blank when creating a new version description the revision select dropdown in the top menu bar in publisher goes blank when creating a new api version current api option is not selected it shows up after page reload img width alt screenshot at src | 0 |
326,615 | 28,006,793,738 | IssuesEvent | 2023-03-27 15:44:35 | primitivefinance/arbiter | https://api.github.com/repos/primitivefinance/arbiter | closed | revm error handling | Needs More Information Testing & Error Handling | We want to be better about our error handling when interacting with revm. We encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until @Autoparallel added the following lines to propagate the error for us.
https://github.com/primitivefinance/arbiter/blob/0e90313fc2caeb31839cdd0d4c4c5a10762563fb/crates/simulate/src/execution.rs#L95-L99 | 1.0 | revm error handling - We want to be better about our error handling when interacting with revm. We encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until @Autoparallel added the following lines to propagate the error for us.
https://github.com/primitivefinance/arbiter/blob/0e90313fc2caeb31839cdd0d4c4c5a10762563fb/crates/simulate/src/execution.rs#L95-L99 | non_defect | revm error handling we want to be better about our error handling when interacting with revm we encountered an error when trying to deploy portfolio regarding the contract size but were unsure why until autoparallel added the following lines to propagate the error for us | 0 |
67,254 | 20,961,596,559 | IssuesEvent | 2022-03-27 21:46:23 | abedmaatalla/imsdroid | https://api.github.com/repos/abedmaatalla/imsdroid | closed | sip register to server in android-ngn-stack | Priority-Medium Type-Defect auto-migrated | ```
when you register to the server use a wrong user information, should get the
`NgnRegistrationEventTypes.REGISTRATION_NOK` event, but actually receive the
`NgnRegistrationEventTypes.UNREGISTRATION_OK` event.
when i check the code in `NgnSipService` class, method `public int
OnDialogEvent(DialogEvent e)` , switch code `case
tinyWRAPConstants.tsip_event_code_dialog_terminated:`, you directed broadcast
event `NgnRegistrationEventTypes.UNREGISTRATION_OK`, and in android-ngn-stack
all code don't have `NgnRegistrationEventTypes.REGISTRATION_NOK` event
broadcast.
```
Original issue reported on code.google.com by `idiottig...@gmail.com` on 8 Feb 2012 at 1:39
| 1.0 | sip register to server in android-ngn-stack - ```
when you register to the server use a wrong user information, should get the
`NgnRegistrationEventTypes.REGISTRATION_NOK` event, but actually receive the
`NgnRegistrationEventTypes.UNREGISTRATION_OK` event.
when i check the code in `NgnSipService` class, method `public int
OnDialogEvent(DialogEvent e)` , switch code `case
tinyWRAPConstants.tsip_event_code_dialog_terminated:`, you directed broadcast
event `NgnRegistrationEventTypes.UNREGISTRATION_OK`, and in android-ngn-stack
all code don't have `NgnRegistrationEventTypes.REGISTRATION_NOK` event
broadcast.
```
Original issue reported on code.google.com by `idiottig...@gmail.com` on 8 Feb 2012 at 1:39
| defect | sip register to server in android ngn stack when you register to the server use a wrong user information should get the ngnregistrationeventtypes registration nok event but actually receive the ngnregistrationeventtypes unregistration ok event when i check the code in ngnsipservice class method public int ondialogevent dialogevent e switch code case tinywrapconstants tsip event code dialog terminated you directed broadcast event ngnregistrationeventtypes unregistration ok and in android ngn stack all code don t have ngnregistrationeventtypes registration nok event broadcast original issue reported on code google com by idiottig gmail com on feb at | 1 |
1,218 | 2,601,760,220 | IssuesEvent | 2015-02-24 00:34:48 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | Memory Leak "fix" will bug if you leave the game instead of restart it | auto-migrated Component-Logic Milestone-Release Priority-Critical Type-Defect Usability | ```
To do.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 25 Jan 2011 at 11:14 | 1.0 | Memory Leak "fix" will bug if you leave the game instead of restart it - ```
To do.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 25 Jan 2011 at 11:14 | defect | memory leak fix will bug if you leave the game instead of restart it to do original issue reported on code google com by aheinerm on jan at | 1 |
47,532 | 13,056,224,755 | IssuesEvent | 2020-07-30 04:02:44 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | I3File does not correctly follow Python iterator interface (Trac #688) | Migrated from Trac dataio defect | The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.
Reference: http://docs.python.org/library/stdtypes.html#iterator-types
In rare cases, this issue can cause unexpected iterator behavior. For example:
```text
1 it = iter(i3file)
2 frame1 = it.next()
3 for frame in it:
4 # attempt to act on second and all subsequent frames...
```
On the first run through the loop, at line 4, `frame` will equal `frame1`.
I'm giving this a low priority, but figured it should be documented.
Migrated from https://code.icecube.wisc.edu/ticket/688
```json
{
"status": "closed",
"changetime": "2014-03-22T04:28:38",
"description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.",
"reporter": "sjackso",
"cc": "",
"resolution": "fixed",
"_ts": "1395462518000000",
"component": "dataio",
"summary": "I3File does not correctly follow Python iterator interface",
"priority": "minor",
"keywords": "I3File I3SequentialFile iterator",
"time": "2012-09-27T19:52:15",
"milestone": "",
"owner": "sjackso",
"type": "defect"
}
```
| 1.0 | I3File does not correctly follow Python iterator interface (Trac #688) - The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.
Reference: http://docs.python.org/library/stdtypes.html#iterator-types
In rare cases, this issue can cause unexpected iterator behavior. For example:
```text
1 it = iter(i3file)
2 frame1 = it.next()
3 for frame in it:
4 # attempt to act on second and all subsequent frames...
```
On the first run through the loop, at line 4, `frame` will equal `frame1`.
I'm giving this a low priority, but figured it should be documented.
Migrated from https://code.icecube.wisc.edu/ticket/688
```json
{
"status": "closed",
"changetime": "2014-03-22T04:28:38",
"description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.",
"reporter": "sjackso",
"cc": "",
"resolution": "fixed",
"_ts": "1395462518000000",
"component": "dataio",
"summary": "I3File does not correctly follow Python iterator interface",
"priority": "minor",
"keywords": "I3File I3SequentialFile iterator",
"time": "2012-09-27T19:52:15",
"milestone": "",
"owner": "sjackso",
"type": "defect"
}
```
| defect | does not correctly follow python iterator interface trac the python binding for an implemented in c as provides methods next and iter to implement the python iterator interface however it does not support the iterator interface correctly because an is both a container and an iterator a correct implementation would provide a separate iterator class that implemented next and iter while itself would only provide iter reference in rare cases this issue can cause unexpected iterator behavior for example text it iter it next for frame in it attempt to act on second and all subsequent frames on the first run through the loop at line frame will equal i m giving this a low priority but figured it should be documented migrated from json status closed changetime description the python binding for an implemented in c as provides methods next and iter to implement the python iterator interface however it does not support the iterator interface correctly because an is both a container and an iterator a correct implementation would provide a separate iterator class that implemented next and iter while itself would only provide iter n nreference rare cases this issue can cause unexpected iterator behavior for example n n it iter it next for frame in it attempt to act on second and all subsequent frames n n non the first run through the loop at line frame will equal n ni m giving this a low priority but figured it should be documented reporter sjackso cc resolution fixed ts component dataio summary does not correctly follow python iterator interface priority minor keywords iterator time milestone owner sjackso type defect | 1 |
238,775 | 26,155,044,117 | IssuesEvent | 2022-12-30 20:00:12 | labsai/EDDI | https://api.github.com/repos/labsai/EDDI | closed | quarkus-smallrye-openapi-2.14.1.Final.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quarkus-smallrye-openapi-2.14.1.Final.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quarkus-smallrye-openapi version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- quarkus-smallrye-openapi-2.14.1.Final.jar (Root Library)
- smallrye-open-api-core-2.3.1.jar
- jackson-dataformat-yaml-2.13.4.jar
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p>
<p>Release Date: 2022-12-01</p>
<p>Fix Resolution: org.yaml:snakeyaml - 1.31</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | quarkus-smallrye-openapi-2.14.1.Final.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quarkus-smallrye-openapi-2.14.1.Final.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quarkus-smallrye-openapi version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- quarkus-smallrye-openapi-2.14.1.Final.jar (Root Library)
- smallrye-open-api-core-2.3.1.jar
- jackson-dataformat-yaml-2.13.4.jar
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-1471">https://nvd.nist.gov/vuln/detail/CVE-2022-1471</a></p>
<p>Release Date: 2022-12-01</p>
<p>Fix Resolution: org.yaml:snakeyaml - 1.31</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_defect | quarkus smallrye openapi final jar vulnerabilities highest severity is autoclosed vulnerable library quarkus smallrye openapi final jar path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar vulnerabilities cve severity cvss dependency type fixed in quarkus smallrye openapi version remediation available high snakeyaml jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy quarkus smallrye openapi final jar root library smallrye open api core jar jackson dataformat yaml jar x snakeyaml jar vulnerable library found in base branch main vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend | 0 |
298,738 | 22,554,518,241 | IssuesEvent | 2022-06-27 09:03:20 | dell/ansible-powerscale | https://api.github.com/repos/dell/ansible-powerscale | closed | Create Network pool documentation is wrong | documentation | @rajendravarma077 @meshuga @anupamaloke @walker2 @dattaarindam
From Documentation, below is the code to create a network pool. But getting below error stating unsupported parameters name
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Unsupported parameters for (dellemc.powerscale.dellemc_powerscale_subnet) module: access_zone, pool Supported parameters include: api_password, api_user, description, gateway_priority, groupnet_name, netmask, new_subnet_name, onefs_host, port_no, state, subnet_name, subnet_params, verify_ssl"}
Do we need to use dellemc_powerscale_subnet or dellemc_powerscale_networkpool.
- name: Create Network Pool
dellemc_powerscale_subnet:
onefs_host: "{{onefs_host}}"
api_user: "{{api_user}}"
api_password: "{{api_password}}"
verify_ssl: "{{verify_ssl}}"
groupnet: "groupnet0"
subnet: "subnet0"
pool: "Test_Pool_2"
access_zone: "system"
state: "present" | 1.0 | Create Network pool documentation is wrong - @rajendravarma077 @meshuga @anupamaloke @walker2 @dattaarindam
From Documentation, below is the code to create a network pool. But getting below error stating unsupported parameters name
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Unsupported parameters for (dellemc.powerscale.dellemc_powerscale_subnet) module: access_zone, pool Supported parameters include: api_password, api_user, description, gateway_priority, groupnet_name, netmask, new_subnet_name, onefs_host, port_no, state, subnet_name, subnet_params, verify_ssl"}
Do we need to use dellemc_powerscale_subnet or dellemc_powerscale_networkpool.
- name: Create Network Pool
dellemc_powerscale_subnet:
onefs_host: "{{onefs_host}}"
api_user: "{{api_user}}"
api_password: "{{api_password}}"
verify_ssl: "{{verify_ssl}}"
groupnet: "groupnet0"
subnet: "subnet0"
pool: "Test_Pool_2"
access_zone: "system"
state: "present" | non_defect | create network pool documentation is wrong meshuga anupamaloke dattaarindam from documentation below is the code to create a network pool but getting below error stating unsupported parameters name fatal failed changed false msg unsupported parameters for dellemc powerscale dellemc powerscale subnet module access zone pool supported parameters include api password api user description gateway priority groupnet name netmask new subnet name onefs host port no state subnet name subnet params verify ssl do we need to use dellemc powerscale subnet or dellemc powerscale networkpool name create network pool dellemc powerscale subnet onefs host onefs host api user api user api password api password verify ssl verify ssl groupnet subnet pool test pool access zone system state present | 0 |
71,912 | 23,850,012,401 | IssuesEvent | 2022-09-06 17:00:48 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | Markdown doesn't work in threaded replies | T-Defect A-Markdown A-Timeline S-Major O-Occasional | ### Steps to reproduce
1. See a threaded message with formatting, e.g. "**message**"
2. Reply to it
### Outcome
#### What did you expect?
> **message**
#### What happened instead?
> \*\*message\*\*
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
1.4.34
### Homeserver
matrix.org
### Will you send logs?
No
### Are you willing to provide a PR?
No | 1.0 | Markdown doesn't work in threaded replies - ### Steps to reproduce
1. See a threaded message with formatting, e.g. "**message**"
2. Reply to it
### Outcome
#### What did you expect?
> **message**
#### What happened instead?
> \*\*message\*\*
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
1.4.34
### Homeserver
matrix.org
### Will you send logs?
No
### Are you willing to provide a PR?
No | defect | markdown doesn t work in threaded replies steps to reproduce see a threaded message with formatting e g message reply to it outcome what did you expect message what happened instead message your phone model no response operating system version no response application version and app store homeserver matrix org will you send logs no are you willing to provide a pr no | 1 |
236,298 | 19,530,561,947 | IssuesEvent | 2021-12-30 15:59:53 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks - useMetricsExplorerData Hook should paginate | failed-test Team:Infra Monitoring UI | A test failed on a tracked branch
```
Error: Timed out in waitForNextUpdate after 1000ms.
at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13)
at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks/use_metrics_explorer_data.test.tsx:96:5)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - 7.x](https://buildkite.com/elastic/kibana-hourly/builds/1617#9b2d480c-5a37-4a17-9a89-2fd88b3fdfbf)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks","test.name":"useMetricsExplorerData Hook should paginate","test.failCount":3}} --> | 1.0 | Failing test: Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks - useMetricsExplorerData Hook should paginate - A test failed on a tracked branch
```
Error: Timed out in waitForNextUpdate after 1000ms.
at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13)
at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks/use_metrics_explorer_data.test.tsx:96:5)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-245032e27b709b39/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - 7.x](https://buildkite.com/elastic/kibana-hourly/builds/1617#9b2d480c-5a37-4a17-9a89-2fd88b3fdfbf)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/infra/public/pages/metrics/metrics_explorer/hooks","test.name":"useMetricsExplorerData Hook should paginate","test.failCount":3}} --> | non_defect | failing test jest tests x pack plugins infra public pages metrics metrics explorer hooks usemetricsexplorerdata hook should paginate a test failed on a tracked branch error timed out in waitfornextupdate after at waitfornextupdate opt local ssd buildkite builds kb elastic kibana hourly kibana node modules testing library react hooks lib core asyncutils js at object opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins infra public pages metrics metrics explorer hooks use metrics explorer data test tsx at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js first failure | 0 |
62,331 | 17,023,899,382 | IssuesEvent | 2021-07-03 04:26:47 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Bounding box formated as double for a result in jsonv2 format | Component: nominatim Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 12.28pm, Sunday, 2nd March 2014]**
When I was searching for san francisco using GNOME Maps I got a crash in the geocode-glib library. It could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings.
You can see the result here: http://nominatim.openstreetmap.org/search.php?q=San%20Francisco,%20Agusan%20Del%20Sur,%20Republic%20of%20the%20Philippines&format=json&addressdetails=1&limit=11
Note that the second place has the bounding box as double and not string. | 1.0 | Bounding box formated as double for a result in jsonv2 format - **[Submitted to the original trac issue database at 12.28pm, Sunday, 2nd March 2014]**
When I was searching for san francisco using GNOME Maps I got a crash in the geocode-glib library. It could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings.
You can see the result here: http://nominatim.openstreetmap.org/search.php?q=San%20Francisco,%20Agusan%20Del%20Sur,%20Republic%20of%20the%20Philippines&format=json&addressdetails=1&limit=11
Note that the second place has the bounding box as double and not string. | defect | bounding box formated as double for a result in format when i was searching for san francisco using gnome maps i got a crash in the geocode glib library it could be traced back to one item in the search result had the boundingbox array elements formated as double and not as strings you can see the result here note that the second place has the bounding box as double and not string | 1 |
36,722 | 8,093,388,716 | IssuesEvent | 2018-08-10 00:31:50 | CenturyLinkCloud/mdw | https://api.github.com/repos/CenturyLinkCloud/mdw | closed | CLI Import deletes non-mdw local assets | defect | When running `mdw update` to import mdw base assets, all local assets beginning with "com/" are deleted. | 1.0 | CLI Import deletes non-mdw local assets - When running `mdw update` to import mdw base assets, all local assets beginning with "com/" are deleted. | defect | cli import deletes non mdw local assets when running mdw update to import mdw base assets all local assets beginning with com are deleted | 1 |
43,553 | 11,747,012,810 | IssuesEvent | 2020-03-12 12:53:44 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | opened | Events listing pagination and sorting is broken | Defect Outreach hub VAMC system | **Describe the defect**
Events pagination is whacked
Event sorting is all over the place for this one
https://staging.va.gov/outreach-and-events/events/
it seems to be date ASC on each page, but not across all 3 pages.
Three pages exist here, but there should only be one:
https://staging.va.gov/pittsburgh-health-care/events/
* A past featured event shows on all pages.
* Pages 2 and 3 are empty except for the featured event
**To Reproduce**
Steps to reproduce the behavior:
1. Go to above URLs and click around
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**

**Desktop (please complete the following information if relevant, or delete):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
## Labels
- [/] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [/] CMS objective (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
| 1.0 | Events listing pagination and sorting is broken - **Describe the defect**
Events pagination is whacked
Event sorting is all over the place for this one
https://staging.va.gov/outreach-and-events/events/
it seems to be date ASC on each page, but not across all 3 pages.
Three pages exist here, but there should only be one:
https://staging.va.gov/pittsburgh-health-care/events/
* A past featured event shows on all pages.
* Pages 2 and 3 are empty except for the featured event
**To Reproduce**
Steps to reproduce the behavior:
1. Go to above URLs and click around
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**

**Desktop (please complete the following information if relevant, or delete):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
## Labels
- [/] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [/] CMS objective (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
| defect | events listing pagination and sorting is broken describe the defect events pagination is whacked event sorting is all over the place for this one it seems to be date asc on each page but not across all pages three pages exist here but there should only be one a past featured event shows on all pages pages and are empty except for the featured event to reproduce steps to reproduce the behavior go to above urls and click around expected behavior a clear and concise description of what you expected to happen screenshots desktop please complete the following information if relevant or delete os browser version additional context add any other context about the problem here labels issue type red defaults to defect cms subsystem green cms practice area blue cms objective orange not needed for bug tickets cms supported product black | 1 |
287,956 | 21,677,925,236 | IssuesEvent | 2022-05-09 00:54:17 | hfang821/u-develop-it-BE | https://api.github.com/repos/hfang821/u-develop-it-BE | closed | Create the voters table | documentation | **User Story**
* As a user, I can request a list of voters alphabetized by last name.
* As a user, I can request a single voter's information.
* As a user, I can create a voter.
* As a user, I can update a voter's email.
* As a user, I can delete a voter | 1.0 | Create the voters table - **User Story**
* As a user, I can request a list of voters alphabetized by last name.
* As a user, I can request a single voter's information.
* As a user, I can create a voter.
* As a user, I can update a voter's email.
* As a user, I can delete a voter | non_defect | create the voters table user story as a user i can request a list of voters alphabetized by last name as a user i can request a single voter s information as a user i can create a voter as a user i can update a voter s email as a user i can delete a voter | 0 |
41,523 | 10,508,465,011 | IssuesEvent | 2019-09-27 08:43:11 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | LikeEscapeStep#escape() also supported by Teradata | C: DB: Teradata C: DB: Teradata Aster E: Enterprise Edition P: Medium R: Fixed T: Defect | The `@Support` annotation on `LikeEscapeStep#escape()` lists most dialects, but `TERADATA` appears to be missing. According to the official documentation the `ESCAPE` clause is supported: https://docs.teradata.com/reader/756LNiPSFdY~4JcCCcR5Cw/ZP3CE_cR~e7V50zVkzzeVQ. | 1.0 | LikeEscapeStep#escape() also supported by Teradata - The `@Support` annotation on `LikeEscapeStep#escape()` lists most dialects, but `TERADATA` appears to be missing. According to the official documentation the `ESCAPE` clause is supported: https://docs.teradata.com/reader/756LNiPSFdY~4JcCCcR5Cw/ZP3CE_cR~e7V50zVkzzeVQ. | defect | likeescapestep escape also supported by teradata the support annotation on likeescapestep escape lists most dialects but teradata appears to be missing according to the official documentation the escape clause is supported | 1 |
20,081 | 3,295,315,199 | IssuesEvent | 2015-10-31 20:48:32 | chief-atx/bcmon | https://api.github.com/repos/chief-atx/bcmon | closed | No development since April 2013 | auto-migrated Priority-Medium Type-Defect | ```
I am forking this project. Is anyone on the planet still interested in
maintaining and/or contributing?
```
Original issue reported on code.google.com by `nethers...@gmail.com` on 20 Mar 2015 at 6:07 | 1.0 | No development since April 2013 - ```
I am forking this project. Is anyone on the planet still interested in
maintaining and/or contributing?
```
Original issue reported on code.google.com by `nethers...@gmail.com` on 20 Mar 2015 at 6:07 | defect | no development since april i am forking this project is anyone on the planet still interested in maintaining and or contributing original issue reported on code google com by nethers gmail com on mar at | 1 |
390,407 | 11,543,316,877 | IssuesEvent | 2020-02-18 09:23:10 | cthit/Gamma | https://api.github.com/repos/cthit/Gamma | opened | Cannot add group post | Priority: Critical Status: Accepted Where: Backend | https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/controller/admin/GroupPostAdminController.java#L61
Tries to see if the swedish name exists, but `postExists` expects an UUID.
https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/service/PostService.java#L44 | 1.0 | Cannot add group post - https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/controller/admin/GroupPostAdminController.java#L61
Tries to see if the swedish name exists, but `postExists` expects an UUID.
https://github.com/cthit/Gamma/blob/9340b24bc19dfcdb71446e27c98b1d5d8956288a/backend/src/main/java/it/chalmers/gamma/service/PostService.java#L44 | non_defect | cannot add group post tries to see if the swedish name exists but postexists expects an uuid | 0 |
53,493 | 13,261,753,733 | IssuesEvent | 2020-08-20 20:28:23 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [topsimulator] has a number of #warnings (Trac #1529) | Migrated from Trac combo simulation defect | These look like TODOs
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1529">https://code.icecube.wisc.edu/projects/icecube/ticket/1529</a>, reported by kjmeagherand owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"_ts": "1550067215093672",
"description": "These look like TODOs",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-01-26T10:20:59",
"component": "combo simulation",
"summary": "[topsimulator] has a number of #warnings",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [topsimulator] has a number of #warnings (Trac #1529) - These look like TODOs
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1529">https://code.icecube.wisc.edu/projects/icecube/ticket/1529</a>, reported by kjmeagherand owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"_ts": "1550067215093672",
"description": "These look like TODOs",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-01-26T10:20:59",
"component": "combo simulation",
"summary": "[topsimulator] has a number of #warnings",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| defect | has a number of warnings trac these look like todos migrated from json status closed changetime ts description these look like todos reporter kjmeagher cc resolution fixed time component combo simulation summary has a number of warnings priority normal keywords milestone owner jgonzalez type defect | 1 |
5,238 | 2,610,183,905 | IssuesEvent | 2015-02-26 18:58:30 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 证实长了色斑该怎么办 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我��
�静静的想着某人;伤心的时候,我会找个地方静静的发呆,�
��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装
自己,对别人说:我很好、我很开心;失落的时候,我会笑��
�对自己说,没事的,一切总会过去。黄褐斑最主要的就是不�
��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事
情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来��
�,这才能有效去除黄褐斑。长了色斑该怎么办,
《客户案例》
因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比��
�了解,现在我要说的是还是精华祛斑比较好,我就是用精华�
��斑的方法祛斑的。<br>
我是因为工作的关系需要经常化妆,大家都知道,化妆��
�都会或多或少的含有一些化学成分,慢慢的,我的脸上就长�
��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道
斑也分很多种,有先天的,也有后天的,先天的一般是很难��
�除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天�
��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法
完全去掉的。刚开始也是受化妆品的影响,用了很多的美白��
�斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又�
��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知
识,才知道那些快速口服药中一般含有重金属,短期效果明��
�,但是一旦停用,就会反复长更多的斑,而且这样的情况是�
��难被改善的,所以我就再不用那些美白祛斑的口服的了。<br
>
我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个��
�纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯�
��华的应该没什么副作用,再说是朋友介绍的,应该是有用,
就去他们商城上详细询问了客服,告诉我这个主要是通过调��
�祛斑的,对身体没任何副作用,我又查了这个产品的各个成�
��的作用,确实还挺不错的,就一下订购了三个周期的,真没
想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,��
�在皮肤也好了很多,也没反弹。
阅读了长了色斑该怎么办,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
长了色斑该怎么办,同时为您分享祛斑小方法
天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙��
�光泽
必须长期内调,做好补血养气的工作才能让自己摆脱天生长��
�的问题。红枣、阿胶、
红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物��
�很好的补气作用。当然除了补气补血外,体内调理好了,那�
��长斑的机会就更少了。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:11 | 1.0 | 证实长了色斑该怎么办 - ```
《摘要》
哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我��
�静静的想着某人;伤心的时候,我会找个地方静静的发呆,�
��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装
自己,对别人说:我很好、我很开心;失落的时候,我会笑��
�对自己说,没事的,一切总会过去。黄褐斑最主要的就是不�
��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事
情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来��
�,这才能有效去除黄褐斑。长了色斑该怎么办,
《客户案例》
因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比��
�了解,现在我要说的是还是精华祛斑比较好,我就是用精华�
��斑的方法祛斑的。<br>
我是因为工作的关系需要经常化妆,大家都知道,化妆��
�都会或多或少的含有一些化学成分,慢慢的,我的脸上就长�
��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道
斑也分很多种,有先天的,也有后天的,先天的一般是很难��
�除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天�
��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法
完全去掉的。刚开始也是受化妆品的影响,用了很多的美白��
�斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又�
��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知
识,才知道那些快速口服药中一般含有重金属,短期效果明��
�,但是一旦停用,就会反复长更多的斑,而且这样的情况是�
��难被改善的,所以我就再不用那些美白祛斑的口服的了。<br
>
我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个��
�纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯�
��华的应该没什么副作用,再说是朋友介绍的,应该是有用,
就去他们商城上详细询问了客服,告诉我这个主要是通过调��
�祛斑的,对身体没任何副作用,我又查了这个产品的各个成�
��的作用,确实还挺不错的,就一下订购了三个周期的,真没
想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,��
�在皮肤也好了很多,也没反弹。
阅读了长了色斑该怎么办,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
长了色斑该怎么办,同时为您分享祛斑小方法
天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙��
�光泽
必须长期内调,做好补血养气的工作才能让自己摆脱天生长��
�的问题。红枣、阿胶、
红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物��
�很好的补气作用。当然除了补气补血外,体内调理好了,那�
��长斑的机会就更少了。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:11 | defect | 证实长了色斑该怎么办 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。长了色斑该怎么办, 《客户案例》 因为长斑了三年多了,对怎么祛斑和一些祛斑方法都比�� �了解,现在我要说的是还是精华祛斑比较好,我就是用精华� ��斑的方法祛斑的。 我是因为工作的关系需要经常化妆,大家都知道,化妆�� �都会或多或少的含有一些化学成分,慢慢的,我的脸上就长� ��一些斑,为了祛斑,我也找了很多这方面的东西看,才知道 斑也分很多种,有先天的,也有后天的,先天的一般是很难�� �除的,就像遗传性黄褐斑,这个就很难去掉了,再就是后天� ��,像晒斑啊辐射斑啊妊娠斑黄褐斑,还有我这种斑就有办法 完全去掉的。刚开始也是受化妆品的影响,用了很多的美白�� �斑的口服药,可那些都是起些暂时的作用,斑去掉没多久又� ��来了,真的挺烦人的,后来慢慢的了解了一些祛斑方面的知 识,才知道那些快速口服药中一般含有重金属,短期效果明�� �,但是一旦停用,就会反复长更多的斑,而且这样的情况是� ��难被改善的,所以我就再不用那些美白祛斑的口服的了。 br 我用「黛芙薇尔精华液」也是通过朋友介绍的,说这个�� �纯精华的,祛斑效果挺好的,还不反弹,我想这个既然是纯� ��华的应该没什么副作用,再说是朋友介绍的,应该是有用, 就去他们商城上详细询问了客服,告诉我这个主要是通过调�� �祛斑的,对身体没任何副作用,我又查了这个产品的各个成� ��的作用,确实还挺不错的,就一下订购了三个周期的,真没 想到,我这次还真找对祛斑产品了,我的斑还真的去掉了,�� �在皮肤也好了很多,也没反弹。 阅读了长了色斑该怎么办,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 长了色斑该怎么办,同时为您分享祛斑小方法 天生长斑最多见于天生脾脏不太好的人群中,想要肌肤白皙�� �光泽 必须长期内调,做好补血养气的工作才能让自己摆脱天生长�� �的问题。红枣、阿胶、 红豆等都是补血的佳品,山药、洋芋、土豆这些常见的食物�� �很好的补气作用。当然除了补气补血外,体内调理好了,那� ��长斑的机会就更少了。 original issue reported on code google com by additive gmail com on jul at | 1 |
68,774 | 21,886,101,546 | IssuesEvent | 2022-05-19 18:48:33 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Excess parentheses generated in unary ~ (QOM.BitNot) or - (QOM.Neg) operators | T: Defect C: Functionality P: Medium E: All Editions | Like https://github.com/jOOQ/jOOQ/issues/7362, we tend to produce too many parentheses around the argument of unary `~` (`QOM.BitNot`) and `-` (`QOM.Neg`) operators. E.g.
Input:
```sql
select
~(1 + 2),
-(1 + 2)
```
Output:
```sql
select
~((1 + 2)),
-((1 + 2))
``` | 1.0 | Excess parentheses generated in unary ~ (QOM.BitNot) or - (QOM.Neg) operators - Like https://github.com/jOOQ/jOOQ/issues/7362, we tend to produce too many parentheses around the argument of unary `~` (`QOM.BitNot`) and `-` (`QOM.Neg`) operators. E.g.
Input:
```sql
select
~(1 + 2),
-(1 + 2)
```
Output:
```sql
select
~((1 + 2)),
-((1 + 2))
``` | defect | excess parentheses generated in unary qom bitnot or qom neg operators like we tend to produce too many parentheses around the argument of unary qom bitnot and qom neg operators e g input sql select output sql select | 1 |
10,850 | 2,622,193,255 | IssuesEvent | 2015-03-04 00:24:01 | byzhang/cudpp | https://api.github.com/repos/byzhang/cudpp | opened | make install omits cudpp_config.h | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. mkdir -p build; cd build;
2. cmake ..
3. make
4. sudo make install
What is the expected output? What do you see instead?
Expect /usr/local/include to contain cudpp_config.h, but it isn't there.
This results in the following error when building a 3rd-party project that depends on cudpp:
/usr/local/include/cudpp.h:501:26: fatal error: cudpp_config.h: No such file or directory
What version of the product are you using? On what operating system?
cudpp 2.0
ubuntu 12.04
cmake 2.8.7
Please provide any additional information below.
```
Original issue reported on code.google.com by `ksi...@email.arizona.edu` on 5 Jun 2012 at 11:12 | 1.0 | make install omits cudpp_config.h - ```
What steps will reproduce the problem?
1. mkdir -p build; cd build;
2. cmake ..
3. make
4. sudo make install
What is the expected output? What do you see instead?
Expect /usr/local/include to contain cudpp_config.h, but it isn't there.
This results in the following error when building a 3rd-party project that depends on cudpp:
/usr/local/include/cudpp.h:501:26: fatal error: cudpp_config.h: No such file or directory
What version of the product are you using? On what operating system?
cudpp 2.0
ubuntu 12.04
cmake 2.8.7
Please provide any additional information below.
```
Original issue reported on code.google.com by `ksi...@email.arizona.edu` on 5 Jun 2012 at 11:12 | defect | make install omits cudpp config h what steps will reproduce the problem mkdir p build cd build cmake make sudo make install what is the expected output what do you see instead expect usr local include to contain cudpp config h but it isn t there this results in the following error when building a party project that depends on cudpp usr local include cudpp h fatal error cudpp config h no such file or directory what version of the product are you using on what operating system cudpp ubuntu cmake please provide any additional information below original issue reported on code google com by ksi email arizona edu on jun at | 1 |
65,226 | 19,277,372,001 | IssuesEvent | 2021-12-10 13:28:22 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Polls appear as empty events when I don't have the Polls lab flag enabled | T-Defect | ### Steps to reproduce
1. Disable Polls in labs
2. As another user with Polls enabled, create a poll
### Outcome
#### What did you expect?
I should see nothing
#### What happened instead?
I see an empty event
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
Element version: ebc2bb7c5644-react-69c847b94c7f-js-dd8c157bb95a Olm version: 3.2.3
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Polls appear as empty events when I don't have the Polls lab flag enabled - ### Steps to reproduce
1. Disable Polls in labs
2. As another user with Polls enabled, create a poll
### Outcome
#### What did you expect?
I should see nothing
#### What happened instead?
I see an empty event
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
Element version: ebc2bb7c5644-react-69c847b94c7f-js-dd8c157bb95a Olm version: 3.2.3
### Homeserver
_No response_
### Will you send logs?
No | defect | polls appear as empty events when i don t have the polls lab flag enabled steps to reproduce disable polls in labs as another user with polls enabled create a poll outcome what did you expect i should see nothing what happened instead i see an empty event operating system no response browser information no response url for webapp no response application version element version react js olm version homeserver no response will you send logs no | 1 |
18,395 | 3,054,484,972 | IssuesEvent | 2015-08-13 03:06:56 | eczarny/spectacle | https://api.github.com/repos/eczarny/spectacle | closed | AutomaticUpdateCheckEnabled not effective | defect ★ | We deploy Spectacle to our fleet of school Macs with OS X 10.9 via munki. Most devices were at Spectacle 0.8.6. We manage all updates via munki; we never want any software to prompts to update.
Our post-install-script includes the following:
`# disable Spectacle autoupdate
defaults write /Library/Preferences/com.divisiblebyzero.Spectacle AutomaticUpdateCheckEnabled -bool false`
However, this morning with the release of 0.8.8, end-users are being prompted to upgrade.
`defaults read /Library/Preferences/com.divisiblebyzero.Spectacle` on a non-upgraded machine (still running 0.8.6) yields, both before and after 0.8.8 is installed via munki:
`{`
` AutomaticUpdateCheckEnabled = 0;`
`}`
Has the "disabling Automatic Update" mechanism changed?
Thanks,
Mike
| 1.0 | AutomaticUpdateCheckEnabled not effective - We deploy Spectacle to our fleet of school Macs with OS X 10.9 via munki. Most devices were at Spectacle 0.8.6. We manage all updates via munki; we never want any software to prompts to update.
Our post-install-script includes the following:
`# disable Spectacle autoupdate
defaults write /Library/Preferences/com.divisiblebyzero.Spectacle AutomaticUpdateCheckEnabled -bool false`
However, this morning with the release of 0.8.8, end-users are being prompted to upgrade.
`defaults read /Library/Preferences/com.divisiblebyzero.Spectacle` on a non-upgraded machine (still running 0.8.6) yields, both before and after 0.8.8 is installed via munki:
`{`
` AutomaticUpdateCheckEnabled = 0;`
`}`
Has the "disabling Automatic Update" mechanism changed?
Thanks,
Mike
| defect | automaticupdatecheckenabled not effective we deploy spectacle to our fleet of school macs with os x via munki most devices were at spectacle we manage all updates via munki we never want any software to prompts to update our post install script includes the following disable spectacle autoupdate defaults write library preferences com divisiblebyzero spectacle automaticupdatecheckenabled bool false however this morning with the release of end users are being prompted to upgrade defaults read library preferences com divisiblebyzero spectacle on a non upgraded machine still running yields both before and after is installed via munki automaticupdatecheckenabled has the disabling automatic update mechanism changed thanks mike | 1 |
12,333 | 2,692,107,692 | IssuesEvent | 2015-04-01 03:53:58 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | opened | Wrong case on filename if outputBy is "NamespacePath" | defect | #### bridge.json
```
{
output: "Bridge\\output",
outputBy: "NamespacePath"
}
```
#### Demo.cs
```
using Bridge;
namespace Demo.Utils
{
public class App
{
public string Test { get; set; }
}
}
```
`demo > utils.js` is expected, but `demo > Utils.js` is created.

| 1.0 | Wrong case on filename if outputBy is "NamespacePath" - #### bridge.json
```
{
output: "Bridge\\output",
outputBy: "NamespacePath"
}
```
#### Demo.cs
```
using Bridge;
namespace Demo.Utils
{
public class App
{
public string Test { get; set; }
}
}
```
`demo > utils.js` is expected, but `demo > Utils.js` is created.

| defect | wrong case on filename if outputby is namespacepath bridge json output bridge output outputby namespacepath demo cs using bridge namespace demo utils public class app public string test get set demo utils js is expected but demo utils js is created | 1 |
13,464 | 2,759,260,009 | IssuesEvent | 2015-04-28 01:44:14 | jsnjycwyz/gogo-tester | https://api.github.com/repos/jsnjycwyz/gogo-tester | closed | 扫描要的时间越来越久了 | auto-migrated Priority-Medium Type-Defect | ```
以前一下就扫除几十个出来,现在需要的时间越来越久了,��
�一个小时了,才扫出十个
```
Original issue reported on code.google.com by `Cloud...@gmail.com` on 20 Apr 2015 at 3:03 | 1.0 | 扫描要的时间越来越久了 - ```
以前一下就扫除几十个出来,现在需要的时间越来越久了,��
�一个小时了,才扫出十个
```
Original issue reported on code.google.com by `Cloud...@gmail.com` on 20 Apr 2015 at 3:03 | defect | 扫描要的时间越来越久了 以前一下就扫除几十个出来,现在需要的时间越来越久了,�� �一个小时了,才扫出十个 original issue reported on code google com by cloud gmail com on apr at | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.