Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
151,803 | 19,665,390,103 | IssuesEvent | 2022-01-10 21:50:02 | rsoreq/grafana | https://api.github.com/repos/rsoreq/grafana | opened | WS-2022-0008 (Medium) detected in node-forge-0.9.0.tgz | security vulnerability | ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /yarn.lock</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","packageFilePaths":["/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"webpack-dev-server:3.2.1;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2022-0008","vulnerabilityDetails":"The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.","vulnerabilityUrl":"https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | WS-2022-0008 (Medium) detected in node-forge-0.9.0.tgz - ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /yarn.lock</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","packageFilePaths":["/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"webpack-dev-server:3.2.1;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2022-0008","vulnerabilityDetails":"The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.","vulnerabilityUrl":"https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_defect | ws medium detected in node forge tgz ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file yarn lock path to vulnerable library node modules node forge package json dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in base branch master vulnerability details the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree webpack dev server selfsigned node forge isminimumfixversionavailable true minimumfixversion node forge isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way vulnerabilityurl | 0 |
64,983 | 18,992,037,257 | IssuesEvent | 2021-11-22 08:35:38 | zotonic/zotonic | https://api.github.com/repos/zotonic/zotonic | closed | Articles changed to Collections do not show Contains relationship until system cache is flushed | defect admin-ui | When changing an article to a collection, the admin interface fails to show the now-relevant _Contains_ relationship. This confuses the user. Flushing system caches is a workaround.
Steps to reproduce:
1. Create new Article
2. Change Article to Collection
3. Observe: no _Contains_ relationship is shown
4. Flush system caches
5. Observe: _Contains_ relationship is now shown | 1.0 | Articles changed to Collections do not show Contains relationship until system cache is flushed - When changing an article to a collection, the admin interface fails to show the now-relevant _Contains_ relationship. This confuses the user. Flushing system caches is a workaround.
Steps to reproduce:
1. Create new Article
2. Change Article to Collection
3. Observe: no _Contains_ relationship is shown
4. Flush system caches
5. Observe: _Contains_ relationship is now shown | defect | articles changed to collections do not show contains relationship until system cache is flushed when changing an article to a collection the admin interface fails to show the now relevant contains relationship this confuses the user flushing system caches is a workaround steps to reproduce create new article change article to collection observe no contains relationship is shown flush system caches observe contains relationship is now shown | 1 |
195,492 | 6,912,379,745 | IssuesEvent | 2017-11-28 11:45:42 | geosolutions-it/unesco-ihp | https://api.github.com/repos/geosolutions-it/unesco-ihp | closed | Print preview extent (GeoExplorer map client) | frontend pending review Priority: High unesco-ihp | This follows issue #60. Since there are still problems, I am creating a dedicated issue for this point. | 1.0 | Print preview extent (GeoExplorer map client) - This follows issue #60. Since there are still problems, I am creating a dedicated issue for this point. | non_defect | print preview extent geoexplorer map client this follows issue since there are still problems i am creating a dedicated issue for this point | 0 |
5,791 | 3,285,571,125 | IssuesEvent | 2015-10-28 21:11:13 | PolymerElements/paper-tabs | https://api.github.com/repos/PolymerElements/paper-tabs | closed | Keyboard navigation for paper-tabs should stop event propagation | code-yellow | This is important when apps want to use global keyboard shortcuts. As is, both the tab navigation action, and the global shortcut action are performed, which is bad for a11y. | 1.0 | Keyboard navigation for paper-tabs should stop event propagation - This is important when apps want to use global keyboard shortcuts. As is, both the tab navigation action, and the global shortcut action are performed, which is bad for a11y. | non_defect | keyboard navigation for paper tabs should stop event propagation this is important when apps want to use global keyboard shortcuts as is both the tab navigation action and the global shortcut action are performed which is bad for | 0 |
7,429 | 2,610,382,644 | IssuesEvent | 2015-02-26 20:03:52 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Using Land Spray causes the game to crash if "More Wind" is toggled on. | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Create a game scheme with "More Wind" option toggled on.
2. Run the game.
3. Try to use Land Spray.
What is the expected output? What do you see instead?
The game crashes to menu.
What version of the product are you using? On what operating system?
0.9.16 on Win XP SP3
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 18 Sep 2011 at 7:39 | 1.0 | Using Land Spray causes the game to crash if "More Wind" is toggled on. - ```
What steps will reproduce the problem?
1. Create a game scheme with "More Wind" option toggled on.
2. Run the game.
3. Try to use Land Spray.
What is the expected output? What do you see instead?
The game crashes to menu.
What version of the product are you using? On what operating system?
0.9.16 on Win XP SP3
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 18 Sep 2011 at 7:39 | defect | using land spray causes the game to crash if more wind is toggled on what steps will reproduce the problem create a game scheme with more wind option toggled on run the game try to use land spray what is the expected output what do you see instead the game crashes to menu what version of the product are you using on what operating system on win xp please provide any additional information below original issue reported on code google com by adibiaz gmail com on sep at | 1 |
76,825 | 26,617,759,173 | IssuesEvent | 2023-01-24 08:51:34 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | (Rich text editor) After typing a message in a thread, keyboard focus moves to the main timeline composer | T-Defect S-Minor O-Occasional A-Threads Z-Labs A-Rich-Text-Editor | ### Steps to reproduce
1. With rich text editor labs flag enabled
2. Open a thread, type a message and press Enter to send
### Outcome
#### What did you expect?
I would expect the keyboard focus to stay in the same composer where I typed my message (in the thread).
#### What happened instead?
The keyboard focus jumped to the main timeline composer.
### Operating system
Ubuntu 22.04
### Browser information
Firefox 106.0.2
### URL for webapp
https://develop.element.io
### Application version
Element version: 65b8bc196a89-react-3f3005a3ca4d-js-bd4589fcc45b Olm version: 3.2.12
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | (Rich text editor) After typing a message in a thread, keyboard focus moves to the main timeline composer - ### Steps to reproduce
1. With rich text editor labs flag enabled
2. Open a thread, type a message and press Enter to send
### Outcome
#### What did you expect?
I would expect the keyboard focus to stay in the same composer where I typed my message (in the thread).
#### What happened instead?
The keyboard focus jumped to the main timeline composer.
### Operating system
Ubuntu 22.04
### Browser information
Firefox 106.0.2
### URL for webapp
https://develop.element.io
### Application version
Element version: 65b8bc196a89-react-3f3005a3ca4d-js-bd4589fcc45b Olm version: 3.2.12
### Homeserver
matrix.org
### Will you send logs?
No | defect | rich text editor after typing a message in a thread keyboard focus moves to the main timeline composer steps to reproduce with rich text editor labs flag enabled open a thread type a message and press enter to send outcome what did you expect i would expect the keyboard focus to stay in the same composer where i typed my message in the thread what happened instead the keyboard focus jumped to the main timeline composer operating system ubuntu browser information firefox url for webapp application version element version react js olm version homeserver matrix org will you send logs no | 1 |
46,251 | 13,055,876,057 | IssuesEvent | 2020-07-30 02:59:54 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | dataclasses does not build with Boost>1.47 (Trac #832) | Incomplete Migration Migrated from Trac combo simulation defect | Migrated from https://code.icecube.wisc.edu/ticket/832
```json
{
"status": "closed",
"changetime": "2015-01-26T22:20:20",
"description": "I am trying to build the trunk version of simulation but cannot get dataclasses to compile. The problem is with I3MCTree.cxx under pybindings. The error is associated with Boost's python function range. I have tried all versions of Boost on our machines (1.47.0, 1.51.0, 1.52.0, 1.55.0, 1.57.0) and the error is the same. The output I get from running cmake is:\n\n{{{\n\n[ 40%] Building CXX object dataclasses/CMakeFiles/dataclasses-pybindings.dir/private/pybindings/I3MCTree.cxx.o\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx: In function \u2018I3Particle* at2(I3MCTree&, int)\u2019:\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:83: warning: comparison between signed and unsigned integer expressions\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:87: warning: comparison between signed and unsigned integer expressions\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp: In function \u2018void register_I3MCTree()\u2019:\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:247: error: in call to \u2018range\u2019\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:262: error: in call to \u2018range\u2019\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:281: error: in call to \u2018range\u2019\nmake[2]: *** [dataclasses/CMakeFiles/dataclasses-pybindings.dir/private/pybindings/I3MCTree.cxx.o] Error 1\nmake[1]: *** [dataclasses/CMakeFiles/dataclasses-pybindings.dir/all] Error 2\nmake: *** [all] Error 2\n\n}}}\n\nThis error only started during the simulation workshop. Before then everything built without any errors. \n\nOther possible useful information:\ncompiler: gcc-4.4.5\ncmake: 2.8.5\nos: RHEL 6.5",
"reporter": "hignight",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"_ts": "1422310820471928",
"component": "combo simulation",
"summary": "dataclasses does not build with Boost>1.47",
"priority": "critical",
"keywords": "dataclasses",
"time": "2014-12-16T19:46:27",
"milestone": "",
"owner": "",
"type": "defect"
}
```
| 1.0 | dataclasses does not build with Boost>1.47 (Trac #832) - Migrated from https://code.icecube.wisc.edu/ticket/832
```json
{
"status": "closed",
"changetime": "2015-01-26T22:20:20",
"description": "I am trying to build the trunk version of simulation but cannot get dataclasses to compile. The problem is with I3MCTree.cxx under pybindings. The error is associated with Boost's python function range. I have tried all versions of Boost on our machines (1.47.0, 1.51.0, 1.52.0, 1.55.0, 1.57.0) and the error is the same. The output I get from running cmake is:\n\n{{{\n\n[ 40%] Building CXX object dataclasses/CMakeFiles/dataclasses-pybindings.dir/private/pybindings/I3MCTree.cxx.o\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx: In function \u2018I3Particle* at2(I3MCTree&, int)\u2019:\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:83: warning: comparison between signed and unsigned integer expressions\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:87: warning: comparison between signed and unsigned integer expressions\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp: In function \u2018void register_I3MCTree()\u2019:\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:247: error: in call to \u2018range\u2019\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:262: error: in call to \u2018range\u2019\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/range/algorithm/equal.hpp:160: error: \u2018boost::range\u2019 is not a function,\n/opt/software/boost/1.55.0--GCC-4.4.5--OpenMPI-1.4.3/include/boost/python/iterator.hpp:112: error: conflict with \u2018template<class NextPolicies, class Target, class Accessor1, class Accessor2> boost::python::api::object boost::python::range(Accessor1, Accessor2, NextPolicies*, boost::type<U>*)\u2019\n/mnt/home/hignight/work/simulation_workshop/trunk/dataclasses/private/pybindings/I3MCTree.cxx:281: error: in call to \u2018range\u2019\nmake[2]: *** [dataclasses/CMakeFiles/dataclasses-pybindings.dir/private/pybindings/I3MCTree.cxx.o] Error 1\nmake[1]: *** [dataclasses/CMakeFiles/dataclasses-pybindings.dir/all] Error 2\nmake: *** [all] Error 2\n\n}}}\n\nThis error only started during the simulation workshop. Before then everything built without any errors. \n\nOther possible useful information:\ncompiler: gcc-4.4.5\ncmake: 2.8.5\nos: RHEL 6.5",
"reporter": "hignight",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"_ts": "1422310820471928",
"component": "combo simulation",
"summary": "dataclasses does not build with Boost>1.47",
"priority": "critical",
"keywords": "dataclasses",
"time": "2014-12-16T19:46:27",
"milestone": "",
"owner": "",
"type": "defect"
}
```
| defect | dataclasses does not build with boost trac migrated from json status closed changetime description i am trying to build the trunk version of simulation but cannot get dataclasses to compile the problem is with cxx under pybindings the error is associated with boost s python function range i have tried all versions of boost on our machines and the error is the same the output i get from running cmake is n n n n building cxx object dataclasses cmakefiles dataclasses pybindings dir private pybindings cxx o n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx in function int n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx warning comparison between signed and unsigned integer expressions n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx warning comparison between signed and unsigned integer expressions n opt software boost gcc openmpi include boost range algorithm equal hpp in function register n opt software boost gcc openmpi include boost range algorithm equal hpp error range is not a function n opt software boost gcc openmpi include boost python iterator hpp error conflict with boost python api object boost python range nextpolicies boost type n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx error in call to n opt software boost gcc openmpi include boost range algorithm equal hpp error range is not a function n opt software boost gcc openmpi include boost python iterator hpp error conflict with boost python api object boost python range nextpolicies boost type n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx error in call to n opt software boost gcc openmpi include boost range algorithm equal hpp error range is not a function n opt software boost gcc openmpi include boost python iterator hpp error conflict with boost python api object boost python range nextpolicies boost type n mnt home hignight work simulation workshop trunk dataclasses private pybindings cxx error in call to nmake error nmake error nmake error n n n nthis error only started during the simulation workshop before then everything built without any errors n nother possible useful information ncompiler gcc ncmake nos rhel reporter hignight cc david schultz icecube wisc edu resolution fixed ts component combo simulation summary dataclasses does not build with boost priority critical keywords dataclasses time milestone owner type defect | 1 |
158,402 | 13,732,210,801 | IssuesEvent | 2020-10-05 03:55:00 | PopStackHack/mVoterAndroid | https://api.github.com/repos/PopStackHack/mVoterAndroid | closed | [Question] How to set up development environment? | documentation | Could you please provide any guideline on how to setup for development build? It would be very helpful for pull requests. | 1.0 | [Question] How to set up development environment? - Could you please provide any guideline on how to setup for development build? It would be very helpful for pull requests. | non_defect | how to set up development environment could you please provide any guideline on how to setup for development build it would be very helpful for pull requests | 0 |
4,573 | 2,610,118,978 | IssuesEvent | 2015-02-26 18:36:56 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | Cannot add a Blog that uses the MetaWeblog API | auto-migrated Milestone-1.4 Priority-Medium Type-Defect | ```
What's the problem?
In Safari, when I try to add my blog, ScribeFire says “Your blog was added
successfully!” but then I just see the “Add a New Blog” button again, and
in fact it didn’t add my blog after all.
I also tried this on Firefox, and the failure mode is slightly different. When
I finish configuration I get a “Method Not Found” error. Checking on my
server, I see that ScribeFire is sending an empty XML-RPC request. All the
expected XML tags inside the RPC request are there, but their content is empty.
So, the server says "method not found" because no method was specified.
What browser are you using?
Safari 5.0, but also Firefox 3.6.8 (see above).
What version of ScribeFire are you running?
1.3.0.0
```
-----
Original issue reported on code.google.com by `mroberts...@gmail.com` on 23 Aug 2010 at 1:00 | 1.0 | Cannot add a Blog that uses the MetaWeblog API - ```
What's the problem?
In Safari, when I try to add my blog, ScribeFire says “Your blog was added
successfully!” but then I just see the “Add a New Blog” button again, and
in fact it didn’t add my blog after all.
I also tried this on Firefox, and the failure mode is slightly different. When
I finish configuration I get a “Method Not Found” error. Checking on my
server, I see that ScribeFire is sending an empty XML-RPC request. All the
expected XML tags inside the RPC request are there, but their content is empty.
So, the server says "method not found" because no method was specified.
What browser are you using?
Safari 5.0, but also Firefox 3.6.8 (see above).
What version of ScribeFire are you running?
1.3.0.0
```
-----
Original issue reported on code.google.com by `mroberts...@gmail.com` on 23 Aug 2010 at 1:00 | defect | cannot add a blog that uses the metaweblog api what s the problem in safari when i try to add my blog scribefire says “your blog was added successfully ” but then i just see the “add a new blog” button again and in fact it didn’t add my blog after all i also tried this on firefox and the failure mode is slightly different when i finish configuration i get a “method not found” error checking on my server i see that scribefire is sending an empty xml rpc request all the expected xml tags inside the rpc request are there but their content is empty so the server says method not found because no method was specified what browser are you using safari but also firefox see above what version of scribefire are you running original issue reported on code google com by mroberts gmail com on aug at | 1 |
54,643 | 13,796,008,996 | IssuesEvent | 2020-10-09 19:03:20 | Reckue/post-api | https://api.github.com/repos/Reckue/post-api | closed | Fix using the cloneable | type:defect | Don't use Cloneable. Use constructor of copy or BeanUtils.cloneBean from Apache Commons:
```java
BeanUtils.cloneBean(original)
```
| 1.0 | Fix using the cloneable - Don't use Cloneable. Use constructor of copy or BeanUtils.cloneBean from Apache Commons:
```java
BeanUtils.cloneBean(original)
```
| defect | fix using the cloneable don t use cloneable use constructor of copy or beanutils clonebean from apache commons java beanutils clonebean original | 1 |
41,447 | 10,470,940,910 | IssuesEvent | 2019-09-23 06:16:28 | melink14/rikaikun | https://api.github.com/repos/melink14/rikaikun | closed | Highlight only while holding a hotkey - new feature suggestion | Priority-Medium Type-Defect auto-migrated | ```
Hello.
It's not very comfortable for me when I moving the mouse cursor over the
website and a lot of popups are shown always. Temporary turning off the
extension is also not very comfortable.
I think it will be very useful and comfortable feature for the user if Kanji
card will pop up only if you highlight Kanji while holding some button pressed.
Optional of course.
```
Original issue reported on code.google.com by `E.A.Guba...@gmail.com` on 7 Jul 2014 at 2:58
| 1.0 | Highlight only while holding a hotkey - new feature suggestion - ```
Hello.
It's not very comfortable for me when I moving the mouse cursor over the
website and a lot of popups are shown always. Temporary turning off the
extension is also not very comfortable.
I think it will be very useful and comfortable feature for the user if Kanji
card will pop up only if you highlight Kanji while holding some button pressed.
Optional of course.
```
Original issue reported on code.google.com by `E.A.Guba...@gmail.com` on 7 Jul 2014 at 2:58
| defect | highlight only while holding a hotkey new feature suggestion hello it s not very comfortable for me when i moving the mouse cursor over the website and a lot of popups are shown always temporary turning off the extension is also not very comfortable i think it will be very useful and comfortable feature for the user if kanji card will pop up only if you highlight kanji while holding some button pressed optional of course original issue reported on code google com by e a guba gmail com on jul at | 1 |
349,331 | 31,793,375,615 | IssuesEvent | 2023-09-13 06:09:07 | UA-1023-TAQC/SpaceToStudyTA | https://api.github.com/repos/UA-1023-TAQC/SpaceToStudyTA | opened | [Guest's home page] Verify that a Guest can see the Space2Study platform benefits at the welcoming block #1069 | issue Guest test case | https://github.com/ita-social-projects/SpaceToStudy-Client/issues/1069#issue-1871176497
# [TC-ID] : Title of the test
### Priority
Priority label
## Description
The description should tell the tester what they’re going to test and include any other pertinent information such as the test environment, test data, and preconditions/assumptions.
### Precondition
Any preconditions that must be met prior to the test being executed.
## Test Steps
| Step No. | Step description | Input data | Expected result |
|-------------|:-------------|:-----------|:-----|
| 1. | what a tester should do | | what a tester should see when they do that |
| 2. | second | | second expected |
## Expected Result
The expected result tells the tester what they should experience as a result of the test steps.
This is how the tester determines if the test case is a “pass” or “fail”.
| 1.0 | [Guest's home page] Verify that a Guest can see the Space2Study platform benefits at the welcoming block #1069 - https://github.com/ita-social-projects/SpaceToStudy-Client/issues/1069#issue-1871176497
# [TC-ID] : Title of the test
### Priority
Priority label
## Description
The description should tell the tester what they’re going to test and include any other pertinent information such as the test environment, test data, and preconditions/assumptions.
### Precondition
Any preconditions that must be met prior to the test being executed.
## Test Steps
| Step No. | Step description | Input data | Expected result |
|-------------|:-------------|:-----------|:-----|
| 1. | what a tester should do | | what a tester should see when they do that |
| 2. | second | | second expected |
## Expected Result
The expected result tells the tester what they should experience as a result of the test steps.
This is how the tester determines if the test case is a “pass” or “fail”.
| non_defect | verify that a guest can see the platform benefits at the welcoming block title of the test priority priority label description the description should tell the tester what they’re going to test and include any other pertinent information such as the test environment test data and preconditions assumptions precondition any preconditions that must be met prior to the test being executed test steps step no step description input data expected result what a tester should do what a tester should see when they do that second second expected expected result the expected result tells the tester what they should experience as a result of the test steps this is how the tester determines if the test case is a “pass” or “fail” | 0 |
39,491 | 9,487,781,364 | IssuesEvent | 2019-04-22 17:50:40 | extnet/Ext.NET | https://api.github.com/repos/extnet/Ext.NET | opened | Mouse wheel in NumberField and SpinnerField inverted between 4.7.x and 4.8.x | 4.x defect review-after-extjs-upgrade sencha | Found: 4.8.1
Ext.NET forum thread: [Mouse wheel in number field has changed behavior](https://forums.ext.net/showthread.php?62640)
The mouse wheel behavior in form fields `Ext.form.field.Number` and "Ext.form.field.Spinner` has been inverted between Ext JS 6.6.0 and 6.7.0. Historically, at least about version 4 (Ext.NET 2), it also worked like it did in 6.6.0, so that was a rather unexpected change between minor releases.
To see the change reflect in Sencha, just check the fiddles included in the documentation pages:
- [6.6.0 Ext.form.field.Number](https://docs.sencha.com/extjs/6.6.0/classic/Ext.form.field.Number.html)
- [6.7.0 Ext.form.field.Number](https://docs.sencha.com/extjs/6.7.0/classic/Ext.form.field.Number.html)
- [6.6.0 Ext.form.field.Spinner](https://docs.sencha.com/extjs/6.6.0/classic/Ext.form.field.Spinner.html)
- [6.7.0 Ext.form.field.Spinner](https://docs.sencha.com/extjs/6.7.0/classic/Ext.form.field.Spinner.html) | 1.0 | Mouse wheel in NumberField and SpinnerField inverted between 4.7.x and 4.8.x - Found: 4.8.1
Ext.NET forum thread: [Mouse wheel in number field has changed behavior](https://forums.ext.net/showthread.php?62640)
The mouse wheel behavior in form fields `Ext.form.field.Number` and "Ext.form.field.Spinner` has been inverted between Ext JS 6.6.0 and 6.7.0. Historically, at least about version 4 (Ext.NET 2), it also worked like it did in 6.6.0, so that was a rather unexpected change between minor releases.
To see the change reflect in Sencha, just check the fiddles included in the documentation pages:
- [6.6.0 Ext.form.field.Number](https://docs.sencha.com/extjs/6.6.0/classic/Ext.form.field.Number.html)
- [6.7.0 Ext.form.field.Number](https://docs.sencha.com/extjs/6.7.0/classic/Ext.form.field.Number.html)
- [6.6.0 Ext.form.field.Spinner](https://docs.sencha.com/extjs/6.6.0/classic/Ext.form.field.Spinner.html)
- [6.7.0 Ext.form.field.Spinner](https://docs.sencha.com/extjs/6.7.0/classic/Ext.form.field.Spinner.html) | defect | mouse wheel in numberfield and spinnerfield inverted between x and x found ext net forum thread the mouse wheel behavior in form fields ext form field number and ext form field spinner has been inverted between ext js and historically at least about version ext net it also worked like it did in so that was a rather unexpected change between minor releases to see the change reflect in sencha just check the fiddles included in the documentation pages | 1 |
10,186 | 2,618,940,545 | IssuesEvent | 2015-03-03 00:03:54 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | The additional alien-skirmish ships are out of balance with the campaign ships | auto-migrated Balance Priority-Medium Type-Defect | ```
The new alien ships (except the Garthog's) available for research and
production in skirmish are not properly balanced with the existing campaign
ships. For example, the dargslan default battleship is immediately available
for production and outperforms the Mk1-Mk2 battleships, making them practically
unnecessary.
How to fix:
- Chain the battleship prerequisites to each other
- Keep the campaign battleships unresearched in skirmish, but make them
available
in the campaign
- pace out the HP/DPS of the new ships with the campaign ships.
```
Original issue reported on code.google.com by `akarn...@gmail.com` on 25 Aug 2014 at 12:47 | 1.0 | The additional alien-skirmish ships are out of balance with the campaign ships - ```
The new alien ships (except the Garthog's) available for research and
production in skirmish are not properly balanced with the existing campaign
ships. For example, the dargslan default battleship is immediately available
for production and outperforms the Mk1-Mk2 battleships, making them practically
unnecessary.
How to fix:
- Chain the battleship prerequisites to each other
- Keep the campaign battleships unresearched in skirmish, but make them
available
in the campaign
- pace out the HP/DPS of the new ships with the campaign ships.
```
Original issue reported on code.google.com by `akarn...@gmail.com` on 25 Aug 2014 at 12:47 | defect | the additional alien skirmish ships are out of balance with the campaign ships the new alien ships except the garthog s available for research and production in skirmish are not properly balanced with the existing campaign ships for example the dargslan default battleship is immediately available for production and outperforms the battleships making them practically unnecessary how to fix chain the battleship prerequisites to each other keep the campaign battleships unresearched in skirmish but make them available in the campaign pace out the hp dps of the new ships with the campaign ships original issue reported on code google com by akarn gmail com on aug at | 1 |
78,675 | 27,705,639,107 | IssuesEvent | 2023-03-14 11:00:26 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Automatic pillification shows black avatars at first | T-Defect S-Minor A-Pills O-Occasional | ### Steps to reproduce
1. Open a room with a pill for a non-member
### Outcome
#### What did you expect?
The default avatar while loading the profile.
#### What happened instead?

### Operating system
Windows 10
### Application version
Element Nightly version: 0.0.1-nightly.2023031301 Olm version: 3.2.12
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No | 1.0 | Automatic pillification shows black avatars at first - ### Steps to reproduce
1. Open a room with a pill for a non-member
### Outcome
#### What did you expect?
The default avatar while loading the profile.
#### What happened instead?

### Operating system
Windows 10
### Application version
Element Nightly version: 0.0.1-nightly.2023031301 Olm version: 3.2.12
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No | defect | automatic pillification shows black avatars at first steps to reproduce open a room with a pill for a non member outcome what did you expect the default avatar while loading the profile what happened instead operating system windows application version element nightly version nightly olm version how did you install the app the internet homeserver io will you send logs no | 1 |
42,840 | 5,539,984,370 | IssuesEvent | 2017-03-22 08:46:33 | geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE | https://api.github.com/repos/geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE | reopened | BgKeSCBXK4l/uwTKy9DUfgx/CyCVItNr6eHPpM1+u8YMAXcVqm+942CMRmanL0sAXBsouXeh8j3s67myybcX/1F0T6sb+4IIFhQvwG7rR7gEvnw+Sy9u0X9m7bRSIFNqvz9wQ34GXHUrWkxW8aOsbq/CX1puGrKoNRC8ahZg5Co= | design | rpMNvuW2yDwEf1+hBzqAgp2Ng59gbvYUuMc/BkW6MLpYqgAol0lXO2qBlSrZk21AqUOOXndIzWsMh9jH8qvbzZ+EjuJJhU1BZWKHAfceoTEz+bT8qDcNcNmxcNC/YvycomcoA7hDPlBf7z4yI9/icWOzFWcY3dxkaypErofvQILWhfYEg8K0VlRGChZqMjqLN/epTZsDwioaZQN/gUba4yl5774OB8aZ5Rj6JzSg78oLgpgD/9o+dpiU7JbnGUq9N/epTZsDwioaZQN/gUba4zf3qU2bA8IqGmUDf4FG2uMdnXzPl64LomwqdmrqPG55KKYNW/XivaInJk/TyB4nMmYthbJN21T3si+1vHMkCOY396lNmwPCKhplA3+BRtrj4I5pdilPtSp196ov462YOwBUys4zA1snDDyn0/dgIg2QVokqOCLkh26P9nkJk4+AN/epTZsDwioaZQN/gUba45TTZqDNiG+bZKo1Heot2Cs396lNmwPCKhplA3+BRtrjN/epTZsDwioaZQN/gUba4yhaYYu89QVqE7hZ0VpYKybD02iMewOiV78tcisSK18EN/epTZsDwioaZQN/gUba42ml2H9Y9QeDci3bJjImlNlyflLHN5HyjiHYiRjzBRXunMPWbmKM384nRd9YvImCtDf3qU2bA8IqGmUDf4FG2uNdeP+IPOOT8UHZ/1nyvPMznZM5Hh6wYqDXpmr+G7gmlOwYREfiROJYcnh0VCbJbHfzV9kRwi8LfWagOj7YwcL7S+hgc2kxm3SkbcEIrlqse+z5m9o43r3dfzhrRvkbq0iU4VKzXI/qeVr7dBdOzrItr151v36cQm9KpgH1BE74BDf3qU2bA8IqGmUDf4FG2uNyK5HJsecP8ackBHzcFBjlNJ1/oBGHD38ksfa3sbf2GJ45ekIGjGrIAiORsfiBDp7q+ONfFAByBS2+nXq95rKH8fck+bSswjThWns9uzbswN7VgsKILij7kIj8C0fs5e7i76x8NKj6yPyx/HKd4934N/epTZsDwioaZQN/gUba42cSlC8Ee9WhYv4l7BCXJHK+SePANQHG88RfyJmXlZ1Su13pT2y9RYhmh4CMW6ZGdZjmbeJvHBHRmTMIpn0E+bRq0v9nX0uZ4G3/KoQICBQOJXXpw9mXKi/N0T816nUeKTf3qU2bA8IqGmUDf4FG2uPWtkpUzu0rkUBkKuL7hhr9B79ne9b1SjC3ziTS6OhpTC7qqzqMPHKj49nso5+yEmNRdM8kRYO6dwOteQZhhhZhG+2xNjsWMW4a0GSOMkMR0Df3qU2bA8IqGmUDf4FG2uNRkAzLHvnKdEYpR2U88zbmmfa61nY+cgZUa+SUWxppkkfw5N/yqz6aOcmnUxhbiydMmFlMcYhTfRs/2TsSkA9tY2egmdEtI+HVJmLJAiSE3Df3qU2bA8IqGmUDf4FG2uM396lNmwPCKhplA3+BRtrjbj8oeN2NILdjCjDPtMN++6Vxsk6CpfACK+ZzHcPMUTopE1VFzDtkoUo1b7cAxVFT791iI5dkoxT3EM6FLYsVFEaf/owi/OJAvGjEuwk4Bi2NHgfn4L1EboShEaZbA6KzuYQslxHjLVQLfcZ542R6GPM5FmMjCYhbe+eJqQ+rkj2mUePNFIxhkrUVo0S5p55Rh/IaS5j/lR9xI0iDVvwrqPzPFxHAZXH2y4LHLvzTqy0396lNmwPCKhplA3+BRtrjeJFv0AufVJ20tqRR6lQ3qlbsVImiFM2zAuwOnFP/g1vvLWS8h5/R1vOTnbDRQ130Puc9uV/yAP0lBERy8wWpKWOmzRZxp3WTur9JR/wgA+caZOUkW6SVvaJg+0PGiixD1xr1PnE88aOKiV5RrPUFKw== | 1.0 | BgKeSCBXK4l/uwTKy9DUfgx/CyCVItNr6eHPpM1+u8YMAXcVqm+942CMRmanL0sAXBsouXeh8j3s67myybcX/1F0T6sb+4IIFhQvwG7rR7gEvnw+Sy9u0X9m7bRSIFNqvz9wQ34GXHUrWkxW8aOsbq/CX1puGrKoNRC8ahZg5Co= - rpMNvuW2yDwEf1+hBzqAgp2Ng59gbvYUuMc/BkW6MLpYqgAol0lXO2qBlSrZk21AqUOOXndIzWsMh9jH8qvbzZ+EjuJJhU1BZWKHAfceoTEz+bT8qDcNcNmxcNC/YvycomcoA7hDPlBf7z4yI9/icWOzFWcY3dxkaypErofvQILWhfYEg8K0VlRGChZqMjqLN/epTZsDwioaZQN/gUba4yl5774OB8aZ5Rj6JzSg78oLgpgD/9o+dpiU7JbnGUq9N/epTZsDwioaZQN/gUba4zf3qU2bA8IqGmUDf4FG2uMdnXzPl64LomwqdmrqPG55KKYNW/XivaInJk/TyB4nMmYthbJN21T3si+1vHMkCOY396lNmwPCKhplA3+BRtrj4I5pdilPtSp196ov462YOwBUys4zA1snDDyn0/dgIg2QVokqOCLkh26P9nkJk4+AN/epTZsDwioaZQN/gUba45TTZqDNiG+bZKo1Heot2Cs396lNmwPCKhplA3+BRtrjN/epTZsDwioaZQN/gUba4yhaYYu89QVqE7hZ0VpYKybD02iMewOiV78tcisSK18EN/epTZsDwioaZQN/gUba42ml2H9Y9QeDci3bJjImlNlyflLHN5HyjiHYiRjzBRXunMPWbmKM384nRd9YvImCtDf3qU2bA8IqGmUDf4FG2uNdeP+IPOOT8UHZ/1nyvPMznZM5Hh6wYqDXpmr+G7gmlOwYREfiROJYcnh0VCbJbHfzV9kRwi8LfWagOj7YwcL7S+hgc2kxm3SkbcEIrlqse+z5m9o43r3dfzhrRvkbq0iU4VKzXI/qeVr7dBdOzrItr151v36cQm9KpgH1BE74BDf3qU2bA8IqGmUDf4FG2uNyK5HJsecP8ackBHzcFBjlNJ1/oBGHD38ksfa3sbf2GJ45ekIGjGrIAiORsfiBDp7q+ONfFAByBS2+nXq95rKH8fck+bSswjThWns9uzbswN7VgsKILij7kIj8C0fs5e7i76x8NKj6yPyx/HKd4934N/epTZsDwioaZQN/gUba42cSlC8Ee9WhYv4l7BCXJHK+SePANQHG88RfyJmXlZ1Su13pT2y9RYhmh4CMW6ZGdZjmbeJvHBHRmTMIpn0E+bRq0v9nX0uZ4G3/KoQICBQOJXXpw9mXKi/N0T816nUeKTf3qU2bA8IqGmUDf4FG2uPWtkpUzu0rkUBkKuL7hhr9B79ne9b1SjC3ziTS6OhpTC7qqzqMPHKj49nso5+yEmNRdM8kRYO6dwOteQZhhhZhG+2xNjsWMW4a0GSOMkMR0Df3qU2bA8IqGmUDf4FG2uNRkAzLHvnKdEYpR2U88zbmmfa61nY+cgZUa+SUWxppkkfw5N/yqz6aOcmnUxhbiydMmFlMcYhTfRs/2TsSkA9tY2egmdEtI+HVJmLJAiSE3Df3qU2bA8IqGmUDf4FG2uM396lNmwPCKhplA3+BRtrjbj8oeN2NILdjCjDPtMN++6Vxsk6CpfACK+ZzHcPMUTopE1VFzDtkoUo1b7cAxVFT791iI5dkoxT3EM6FLYsVFEaf/owi/OJAvGjEuwk4Bi2NHgfn4L1EboShEaZbA6KzuYQslxHjLVQLfcZ542R6GPM5FmMjCYhbe+eJqQ+rkj2mUePNFIxhkrUVo0S5p55Rh/IaS5j/lR9xI0iDVvwrqPzPFxHAZXH2y4LHLvzTqy0396lNmwPCKhplA3+BRtrjeJFv0AufVJ20tqRR6lQ3qlbsVImiFM2zAuwOnFP/g1vvLWS8h5/R1vOTnbDRQ130Puc9uV/yAP0lBERy8wWpKWOmzRZxp3WTur9JR/wgA+caZOUkW6SVvaJg+0PGiixD1xr1PnE88aOKiV5RrPUFKw== | non_defect | eptzsdwioazqn eptzsdwioazqn xivainjk an eptzsdwioazqn brtrjn eptzsdwioazqn eptzsdwioazqn eptzsdwioazqn cgzua owi ejqq wga | 0 |
166,666 | 26,393,201,816 | IssuesEvent | 2023-01-12 17:11:05 | regen-network/regen-registry | https://api.github.com/repos/regen-network/regen-registry | closed | Add regen network logo and share buttons to purchase, send, retire confirmation screens | design | Marketing request to get more brand recognition and help users to share more about their purchases. | 1.0 | Add regen network logo and share buttons to purchase, send, retire confirmation screens - Marketing request to get more brand recognition and help users to share more about their purchases. | non_defect | add regen network logo and share buttons to purchase send retire confirmation screens marketing request to get more brand recognition and help users to share more about their purchases | 0 |
52,734 | 13,224,987,420 | IssuesEvent | 2020-08-17 20:15:38 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Nuke TWR (Trac #257) | Migrated from Trac combo core defect | Kill it. Kill it with fire.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/257">https://code.icecube.wisc.edu/projects/icecube/ticket/257</a>, reported by negaand owned by nwhitehorn</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:22:56",
"_ts": "1547241776942948",
"description": "Kill it. Kill it with fire.",
"reporter": "nega",
"cc": "blaufuss, olivas",
"resolution": "fixed",
"time": "2011-05-11T20:21:18",
"component": "combo core",
"summary": "Nuke TWR",
"priority": "normal",
"keywords": "twr",
"milestone": "",
"owner": "nwhitehorn",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Nuke TWR (Trac #257) - Kill it. Kill it with fire.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/257">https://code.icecube.wisc.edu/projects/icecube/ticket/257</a>, reported by negaand owned by nwhitehorn</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T21:22:56",
"_ts": "1547241776942948",
"description": "Kill it. Kill it with fire.",
"reporter": "nega",
"cc": "blaufuss, olivas",
"resolution": "fixed",
"time": "2011-05-11T20:21:18",
"component": "combo core",
"summary": "Nuke TWR",
"priority": "normal",
"keywords": "twr",
"milestone": "",
"owner": "nwhitehorn",
"type": "defect"
}
```
</p>
</details>
| defect | nuke twr trac kill it kill it with fire migrated from json status closed changetime ts description kill it kill it with fire reporter nega cc blaufuss olivas resolution fixed time component combo core summary nuke twr priority normal keywords twr milestone owner nwhitehorn type defect | 1 |
786,807 | 27,694,215,110 | IssuesEvent | 2023-03-14 00:00:02 | zulip/zulip | https://api.github.com/repos/zulip/zulip | opened | Add scheduled data deletion option when deactivating an organization | help wanted area: settings (admin/org) priority: high area: popovers | At present, when an organization is deactivated, there is no automated data deletion process (though data can, of course, be
deleted manually). We should make it possible to schedule automatic data deletion during the deactivation process.
## Server-level configuration
We should add a Django server-level setting for the max/min duration allowed for the data deletion delay.
## UI changes
In the Deactivate organization modal, we should add the following dropdown option:
"Automatically delete all data after"
Options: All of the following as long as they are within the range specified by the server settings: 0 days (immediately), 14 days, 30 days, 90 days, 1 year, Never delete, Custom [default TBD]
The UI should be closely modeled after the UI for invitation expirations:


"Expires on" should be replaced with "Data to be deleted on".
We can try placing this option just above the "Are you sure you want to deactivate this organization?" line.
## Technical details
* Add a new field on `realm`, `scheduled_deletion_date`, which is usually `None` but will now be set during the realm deactivation process.
* Add the data deletion date as a new required API field for organization deactivations.
(A spec should probably define what options we want to offer, how this should work for self-hosted servers where the policies we have on Zulip Cloud may not apply, etc. -- maybe a Django server-level setting for the max/min duration that filters the set of options?)
Follow-ups:
- [ ] Document the new option on `/help/deactivate-your-organization`. | 1.0 | Add scheduled data deletion option when deactivating an organization - At present, when an organization is deactivated, there is no automated data deletion process (though data can, of course, be
deleted manually). We should make it possible to schedule automatic data deletion during the deactivation process.
## Server-level configuration
We should add a Django server-level setting for the max/min duration allowed for the data deletion delay.
## UI changes
In the Deactivate organization modal, we should add the following dropdown option:
"Automatically delete all data after"
Options: All of the following as long as they are within the range specified by the server settings: 0 days (immediately), 14 days, 30 days, 90 days, 1 year, Never delete, Custom [default TBD]
The UI should be closely modeled after the UI for invitation expirations:


"Expires on" should be replaced with "Data to be deleted on".
We can try placing this option just above the "Are you sure you want to deactivate this organization?" line.
## Technical details
* Add a new field on `realm`, `scheduled_deletion_date`, which is usually `None` but will now be set during the realm deactivation process.
* Add the data deletion date as a new required API field for organization deactivations.
(A spec should probably define what options we want to offer, how this should work for self-hosted servers where the policies we have on Zulip Cloud may not apply, etc. -- maybe a Django server-level setting for the max/min duration that filters the set of options?)
Follow-ups:
- [ ] Document the new option on `/help/deactivate-your-organization`. | non_defect | add scheduled data deletion option when deactivating an organization at present when an organization is deactivated there is no automated data deletion process though data can of course be deleted manually we should make it possible to schedule automatic data deletion during the deactivation process server level configuration we should add a django server level setting for the max min duration allowed for the data deletion delay ui changes in the deactivate organization modal we should add the following dropdown option automatically delete all data after options all of the following as long as they are within the range specified by the server settings days immediately days days days year never delete custom the ui should be closely modeled after the ui for invitation expirations expires on should be replaced with data to be deleted on we can try placing this option just above the are you sure you want to deactivate this organization line technical details add a new field on realm scheduled deletion date which is usually none but will now be set during the realm deactivation process add the data deletion date as a new required api field for organization deactivations a spec should probably define what options we want to offer how this should work for self hosted servers where the policies we have on zulip cloud may not apply etc maybe a django server level setting for the max min duration that filters the set of options follow ups document the new option on help deactivate your organization | 0 |
346,275 | 10,410,246,914 | IssuesEvent | 2019-09-13 10:51:14 | code-ready/crc | https://api.github.com/repos/code-ready/crc | closed | Remove curl from integration tests by using Go's `net/http` module | kind/bug priority/minor size/S | ### General information
To make (integration) test steps platform independent, it is desirable to remove calls to `curl` when testing reachability. Maybe use Go's `net/http` module instead.
## CRC version
```bash
OpenShift version: 4.1.11 (not embedded in binary)
```
| 1.0 | Remove curl from integration tests by using Go's `net/http` module - ### General information
To make (integration) test steps platform independent, it is desirable to remove calls to `curl` when testing reachability. Maybe use Go's `net/http` module instead.
## CRC version
```bash
OpenShift version: 4.1.11 (not embedded in binary)
```
| non_defect | remove curl from integration tests by using go s net http module general information to make integration test steps platform independent it is desirable to remove calls to curl when testing reachability maybe use go s net http module instead crc version bash openshift version not embedded in binary | 0 |
42,696 | 11,219,936,066 | IssuesEvent | 2020-01-07 14:52:55 | mozilla-lockwise/lockwise-android | https://api.github.com/repos/mozilla-lockwise/lockwise-android | opened | Discard changes actually keeps the changes done if Save button has been pressed before | defect | ## Steps to reproduce
Open a login to edit it
Change the username value
Tap on Save button
Tap on Close button
Tap on Discard changes and Discard in the confirm dialog
### Expected behavior
Changes are not applied
### Actual behavior
Even though Discard option is selected, changes in username are not discarded
### Device & build information
* Device: Pixel 2
* Build version: master 0dfe1a1
### Notes
Attachments: Please see video: https://youtu.be/QkN9XqJVDro
| 1.0 | Discard changes actually keeps the changes done if Save button has been pressed before - ## Steps to reproduce
Open a login to edit it
Change the username value
Tap on Save button
Tap on Close button
Tap on Discard changes and Discard in the confirm dialog
### Expected behavior
Changes are not applied
### Actual behavior
Even though Discard option is selected, changes in username are not discarded
### Device & build information
* Device: Pixel 2
* Build version: master 0dfe1a1
### Notes
Attachments: Please see video: https://youtu.be/QkN9XqJVDro
| defect | discard changes actually keeps the changes done if save button has been pressed before steps to reproduce open a login to edit it change the username value tap on save button tap on close button tap on discard changes and discard in the confirm dialog expected behavior changes are not applied actual behavior even though discard option is selected changes in username are not discarded device build information device pixel build version master notes attachments please see video | 1 |
81,491 | 30,879,694,033 | IssuesEvent | 2023-08-03 16:36:28 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [🐛 Bug]: mypy error for action_chains | E-easy I-defect | ### What happened?
mypy (1.4.1) is erroring on a incompatible type for `ActionChains(driver)`
As far as I can tell `ActionChains` should allow for all of the `WebDriver` subclasses not just a subset, shouldn't it?
### How can we reproduce the issue?
```shell
from selenium.webdriver.remote.webdriver import WebDriver
driver: WebDriver
ActionChains(driver)
```
### Relevant log output
```shell
Argument 1 to "ActionChains" has incompatible type "selenium.webdriver.remote.webdriver.WebDriver"; expected "selenium.webdriver.chrome.webdriver.WebDriver | selenium.webdriver.firefox.webdriver.WebDriver | selenium.webdriver.safari.webdriver.WebDriver | selenium.webdriver.ie.webdriver.WebDriver | selenium.webdriver.edge.webdriver.WebDriver" [arg-type]
```
```
### Operating System
macOS
### Selenium version
4.11.2
### What are the browser(s) and version(s) where you see this issue?
n/a
### What are the browser driver(s) and version(s) where you see this issue?
n/a
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: mypy error for action_chains - ### What happened?
mypy (1.4.1) is erroring on a incompatible type for `ActionChains(driver)`
As far as I can tell `ActionChains` should allow for all of the `WebDriver` subclasses not just a subset, shouldn't it?
### How can we reproduce the issue?
```shell
from selenium.webdriver.remote.webdriver import WebDriver
driver: WebDriver
ActionChains(driver)
```
### Relevant log output
```shell
Argument 1 to "ActionChains" has incompatible type "selenium.webdriver.remote.webdriver.WebDriver"; expected "selenium.webdriver.chrome.webdriver.WebDriver | selenium.webdriver.firefox.webdriver.WebDriver | selenium.webdriver.safari.webdriver.WebDriver | selenium.webdriver.ie.webdriver.WebDriver | selenium.webdriver.edge.webdriver.WebDriver" [arg-type]
```
```
### Operating System
macOS
### Selenium version
4.11.2
### What are the browser(s) and version(s) where you see this issue?
n/a
### What are the browser driver(s) and version(s) where you see this issue?
n/a
### Are you using Selenium Grid?
_No response_ | defect | mypy error for action chains what happened mypy is erroring on a incompatible type for actionchains driver as far as i can tell actionchains should allow for all of the webdriver subclasses not just a subset shouldn t it how can we reproduce the issue shell from selenium webdriver remote webdriver import webdriver driver webdriver actionchains driver relevant log output shell argument to actionchains has incompatible type selenium webdriver remote webdriver webdriver expected selenium webdriver chrome webdriver webdriver selenium webdriver firefox webdriver webdriver selenium webdriver safari webdriver webdriver selenium webdriver ie webdriver webdriver selenium webdriver edge webdriver webdriver operating system macos selenium version what are the browser s and version s where you see this issue n a what are the browser driver s and version s where you see this issue n a are you using selenium grid no response | 1 |
267,894 | 20,249,950,427 | IssuesEvent | 2022-02-14 16:56:01 | openssl/openssl | https://api.github.com/repos/openssl/openssl | opened | Enc command documentation not accessible | issue: documentation | The enc command documentation is not accessible inline for the master and 3.0 branches as for 1.1.1 or 1.0.2 branches.
They link to a general cmd help instead.
[https://www.openssl.org/docs/man1.1.1/man1/enc.html](url)
[https://www.openssl.org/docs/man3.0/man1/enc.html](url) | 1.0 | Enc command documentation not accessible - The enc command documentation is not accessible inline for the master and 3.0 branches as for 1.1.1 or 1.0.2 branches.
They link to a general cmd help instead.
[https://www.openssl.org/docs/man1.1.1/man1/enc.html](url)
[https://www.openssl.org/docs/man3.0/man1/enc.html](url) | non_defect | enc command documentation not accessible the enc command documentation is not accessible inline for the master and branches as for or branches they link to a general cmd help instead url url | 0 |
87,559 | 15,785,442,247 | IssuesEvent | 2021-04-01 16:20:37 | intuitivelabs/sipcmbeat | https://api.github.com/repos/intuitivelabs/sipcmbeat | closed | unencrypted ip leaks in rate.key | bug security sipcmbeat | From intuitive-issues created by [poandrei](https://github.com/poandrei): intuitivelabs/intuitive-issues#228
example:
```
"rate": {
"key": "reg-new:159.65.252.186",
"exceeded": 0,
"ex_diff": 0,
"crt": 1,
"lim": 20,
"period": 1000,
"since": "2021-03-04T20:43:23.503Z"
},
``` | True | unencrypted ip leaks in rate.key - From intuitive-issues created by [poandrei](https://github.com/poandrei): intuitivelabs/intuitive-issues#228
example:
```
"rate": {
"key": "reg-new:159.65.252.186",
"exceeded": 0,
"ex_diff": 0,
"crt": 1,
"lim": 20,
"period": 1000,
"since": "2021-03-04T20:43:23.503Z"
},
``` | non_defect | unencrypted ip leaks in rate key from intuitive issues created by intuitivelabs intuitive issues example rate key reg new exceeded ex diff crt lim period since | 0 |
108,418 | 9,306,755,338 | IssuesEvent | 2019-03-25 10:29:27 | NickBurneConsulting-GivePanel/givepanel | https://api.github.com/repos/NickBurneConsulting-GivePanel/givepanel | closed | Remove live chat (helpscout) from data capture form | Priority Ready for test | Need to take off live chat (we know about this) | 1.0 | Remove live chat (helpscout) from data capture form - Need to take off live chat (we know about this) | non_defect | remove live chat helpscout from data capture form need to take off live chat we know about this | 0 |
2,264 | 2,603,992,020 | IssuesEvent | 2015-02-24 19:06:52 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳生殖疱疹的症状及治疗 | auto-migrated Priority-Medium Type-Defect | ```
沈阳生殖疱疹的症状及治疗〓沈陽軍區政治部醫院性病〓TEL��
�024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治�
��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共�
��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、
保健、醫療、科研康復為一體的綜合性醫院。是國家首批公��
�甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大�
��、東南大學等知名高等院校的教學醫院。曾被中國人民解放
軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立��
�體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:35 | 1.0 | 沈阳生殖疱疹的症状及治疗 - ```
沈阳生殖疱疹的症状及治疗〓沈陽軍區政治部醫院性病〓TEL��
�024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治�
��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共�
��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、
保健、醫療、科研康復為一體的綜合性醫院。是國家首批公��
�甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大�
��、東南大學等知名高等院校的教學醫院。曾被中國人民解放
軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立��
�體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:35 | defect | 沈阳生殖疱疹的症状及治疗 沈阳生殖疱疹的症状及治疗〓沈陽軍區政治部醫院性病〓tel�� � 〓 , � ��。 。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 original issue reported on code google com by gmail com on jun at | 1 |
123,088 | 16,440,615,998 | IssuesEvent | 2021-05-20 14:00:41 | carbon-design-system/carbon-addons-iot-react | https://api.github.com/repos/carbon-design-system/carbon-addons-iot-react | closed | [Page header] Design specs | design reconciliation package: angular status: needs priority :inbox_tray: status: needs triage :mag: type: contribution :gift: type: discussion :speech_balloon: type: enhancement :bulb: | ### What package is this for?
- [ ] React
- [x] Angular
[Page header - spec sheet.pdf](https://github.com/carbon-design-system/carbon-addons-iot-react/files/6227340/Page.header.-.spec.sheet.pdf)

### Available extra resources
Component guidance reference:
https://pages.github.ibm.com/ai-applications/design/components/page-header/usage
| 1.0 | [Page header] Design specs - ### What package is this for?
- [ ] React
- [x] Angular
[Page header - spec sheet.pdf](https://github.com/carbon-design-system/carbon-addons-iot-react/files/6227340/Page.header.-.spec.sheet.pdf)

### Available extra resources
Component guidance reference:
https://pages.github.ibm.com/ai-applications/design/components/page-header/usage
| non_defect | design specs what package is this for react angular available extra resources component guidance reference | 0 |
222,230 | 24,692,455,254 | IssuesEvent | 2022-10-19 09:32:30 | billmcchesney1/jazz | https://api.github.com/repos/billmcchesney1/jazz | closed | WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed | security vulnerability | ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /core/jazz_ui/package.json</p>
<p>Path to vulnerable library: /core/jazz_ui/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/jsdom/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/acorn-dynamic-import/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/webpack/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- jest-environment-jsdom-23.4.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/jazz/commit/712665b267203375ee4b15e1f8d1ebe08abc1547">712665b267203375ee4b15e1f8d1ebe08abc1547</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (react-scripts): 4.0.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /core/jazz_ui/package.json</p>
<p>Path to vulnerable library: /core/jazz_ui/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/jsdom/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/acorn-dynamic-import/node_modules/acorn/package.json,/templates/react-website-template/app/node_modules/webpack/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- jest-environment-jsdom-23.4.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/jazz/commit/712665b267203375ee4b15e1f8d1ebe08abc1547">712665b267203375ee4b15e1f8d1ebe08abc1547</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (react-scripts): 4.0.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_defect | ws high detected in acorn tgz autoclosed ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file core jazz ui package json path to vulnerable library core jazz ui node modules acorn package json templates react website template app node modules jsdom node modules acorn package json templates react website template app node modules acorn dynamic import node modules acorn package json templates react website template app node modules webpack node modules acorn package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library found in head commit a href found in base branch develop vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution acorn direct dependency fix resolution react scripts check this box to open an automated fix pr | 0 |
2,074 | 2,603,975,894 | IssuesEvent | 2015-02-24 19:01:29 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳阴茎疱疹是怎么治疗 | auto-migrated Priority-Medium Type-Defect | ```
沈阳阴茎疱疹是怎么治疗〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:18 | 1.0 | 沈阳阴茎疱疹是怎么治疗 - ```
沈阳阴茎疱疹是怎么治疗〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:18 | defect | 沈阳阴茎疱疹是怎么治疗 沈阳阴茎疱疹是怎么治疗〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at | 1 |
646,638 | 21,054,624,360 | IssuesEvent | 2022-04-01 01:01:01 | trimble-oss/modus-web-components | https://api.github.com/repos/trimble-oss/modus-web-components | closed | List Sizes | 3 priority:high story | As a user of the design system, I want to have a middle option of size for lists and content tree available in Web Components to meet the needs of my product.
From Ewa below: We just added 40px as a default with 32px small and 48px large.
Acceptance Criteria
Modus List Items now have three size options:
condensed: 32px height
standard: 40px height
large: 48px height
See https://modus.trimble.com/components/lists/#specifications for specs and demo of how they should look
| 1.0 | List Sizes - As a user of the design system, I want to have a middle option of size for lists and content tree available in Web Components to meet the needs of my product.
From Ewa below: We just added 40px as a default with 32px small and 48px large.
Acceptance Criteria
Modus List Items now have three size options:
condensed: 32px height
standard: 40px height
large: 48px height
See https://modus.trimble.com/components/lists/#specifications for specs and demo of how they should look
| non_defect | list sizes as a user of the design system i want to have a middle option of size for lists and content tree available in web components to meet the needs of my product from ewa below we just added as a default with small and large acceptance criteria modus list items now have three size options condensed height standard height large height see for specs and demo of how they should look | 0 |
66,167 | 7,967,138,727 | IssuesEvent | 2018-07-15 10:17:03 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Edit "Publish Date" quickly with input box | Needs Design Feedback [Component] Post Settings [Type] Enhancement | Problem:
Some times we need to publish posts from distance past, 2012 for example. Current way of moving publish date month by month require 72 click to move to 2012, which is not convenient.
Suggested solution:
Let us type the year of publish in a input box, similar to that of "hour" and "minute".
Suggested alternative solution:
Add button to switch each decimal of publish year from 0~9, if input box is not desired. | 1.0 | Edit "Publish Date" quickly with input box - Problem:
Some times we need to publish posts from distance past, 2012 for example. Current way of moving publish date month by month require 72 click to move to 2012, which is not convenient.
Suggested solution:
Let us type the year of publish in a input box, similar to that of "hour" and "minute".
Suggested alternative solution:
Add button to switch each decimal of publish year from 0~9, if input box is not desired. | non_defect | edit publish date quickly with input box problem some times we need to publish posts from distance past for example current way of moving publish date month by month require click to move to which is not convenient suggested solution let us type the year of publish in a input box similar to that of hour and minute suggested alternative solution add button to switch each decimal of publish year from if input box is not desired | 0 |
1,160 | 2,599,418,135 | IssuesEvent | 2015-02-23 08:19:08 | mlively/Phake | https://api.github.com/repos/mlively/Phake | closed | Problem invoking a method named "eval" on a mock | Defect | First things first, thanks for the hard work you put into `Phake`, it's great!
As for the issue, I'm trying to mock [`phpredis`](https://github.com/nicolasff/phpredis) and then use `verify`/`when` on it's [`eval`](https://github.com/nicolasff/phpredis#eval) method, which leads to some odd behaviour.
I expected it to allow me to `verify`/`when` like I can with it's other methods, but instead it just throws the following exception any time `eval` is called on the mock:
```
RedisException: Redis server went away
```
Which is very strange as that's coming from the actual `phpredis` object, which I haven't instantiated.
I'm creating the mock like this:
```php
$this->redis = Phake::mock('redis');
```
I did dig around the `Phake` codebase and found `eval` in `reservedWords` inside the [`MockClass`](https://github.com/mlively/Phake/blob/master/src/Phake/ClassGenerator/MockClass.php#L109) which is no doubt the issue.
As a quick test, I removed `eval` from the `reservedWords` array, but then invoking `eval` on the mock resulted in this:
```
Phake_ClassGenerator_MockClass::instantiate(): The script tried to execute a method or access a property of an incomplete object. Please ensure that the class definition "redis_PHAKE547e744bc8771" of the object you are trying to operate on was loaded _before_ unserialize() gets called or provide a __autoload() function to load the class definition
```
Is there a work around for this situation or could the code be updated to handle this? | 1.0 | Problem invoking a method named "eval" on a mock - First things first, thanks for the hard work you put into `Phake`, it's great!
As for the issue, I'm trying to mock [`phpredis`](https://github.com/nicolasff/phpredis) and then use `verify`/`when` on it's [`eval`](https://github.com/nicolasff/phpredis#eval) method, which leads to some odd behaviour.
I expected it to allow me to `verify`/`when` like I can with it's other methods, but instead it just throws the following exception any time `eval` is called on the mock:
```
RedisException: Redis server went away
```
Which is very strange as that's coming from the actual `phpredis` object, which I haven't instantiated.
I'm creating the mock like this:
```php
$this->redis = Phake::mock('redis');
```
I did dig around the `Phake` codebase and found `eval` in `reservedWords` inside the [`MockClass`](https://github.com/mlively/Phake/blob/master/src/Phake/ClassGenerator/MockClass.php#L109) which is no doubt the issue.
As a quick test, I removed `eval` from the `reservedWords` array, but then invoking `eval` on the mock resulted in this:
```
Phake_ClassGenerator_MockClass::instantiate(): The script tried to execute a method or access a property of an incomplete object. Please ensure that the class definition "redis_PHAKE547e744bc8771" of the object you are trying to operate on was loaded _before_ unserialize() gets called or provide a __autoload() function to load the class definition
```
Is there a work around for this situation or could the code be updated to handle this? | defect | problem invoking a method named eval on a mock first things first thanks for the hard work you put into phake it s great as for the issue i m trying to mock and then use verify when on it s method which leads to some odd behaviour i expected it to allow me to verify when like i can with it s other methods but instead it just throws the following exception any time eval is called on the mock redisexception redis server went away which is very strange as that s coming from the actual phpredis object which i haven t instantiated i m creating the mock like this php this redis phake mock redis i did dig around the phake codebase and found eval in reservedwords inside the which is no doubt the issue as a quick test i removed eval from the reservedwords array but then invoking eval on the mock resulted in this phake classgenerator mockclass instantiate the script tried to execute a method or access a property of an incomplete object please ensure that the class definition redis of the object you are trying to operate on was loaded before unserialize gets called or provide a autoload function to load the class definition is there a work around for this situation or could the code be updated to handle this | 1 |
14,460 | 2,812,669,336 | IssuesEvent | 2015-05-18 10:10:46 | jliljebl/flowblade | https://api.github.com/repos/jliljebl/flowblade | closed | flowblade hanging | auto-migrated Priority-Medium Type-Defect | ```
all details in the attached text file
```
Original issue reported on code.google.com by `alhoussainy` on 30 Dec 2014 at 3:55
Attachments:
* [issue](https://storage.googleapis.com/google-code-attachments/flowblade/issue-101/comment-0/issue)
| 1.0 | flowblade hanging - ```
all details in the attached text file
```
Original issue reported on code.google.com by `alhoussainy` on 30 Dec 2014 at 3:55
Attachments:
* [issue](https://storage.googleapis.com/google-code-attachments/flowblade/issue-101/comment-0/issue)
| defect | flowblade hanging all details in the attached text file original issue reported on code google com by alhoussainy on dec at attachments | 1 |
442,044 | 12,736,797,318 | IssuesEvent | 2020-06-25 17:33:47 | googleapis/google-cloud-cpp | https://api.github.com/repos/googleapis/google-cloud-cpp | closed | heap-use-after-free in CompletionQueueTest.ShutdownWithReschedulingTimer | priority: p2 type: bug | Are we holding the `grpc::Alarm` incorrectly?
[cloud-cpp/github/google-cloud-cpp/master/docker/tsan-presubmit](https://source.cloud.google.com/results/invocations/a5a82acb-c556-4c9a-9cdb-02ba5bd9fef5/targets/cloud-cpp%2Fgithub%2Fgoogle-cloud-cpp%2Fmaster%2Fdocker%2Ftsan-presubmit/log) failure:
```
[ RUN ] CompletionQueueTest.ShutdownWithReschedulingTimer
==================
WARNING: ThreadSanitizer: heap-use-after-free (pid=4743)
Read of size 8 at 0x7b4400001bc8 by thread T3:
#0 grpc_core::ExecCtx::Flush() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/exec_ctx.cc:150 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xd847a)
#1 run_some_timers() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:134 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf97d2)
#2 timer_main_loop() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:237 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf9593)
#3 timer_thread(void*) /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:284 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf94d7)
#4 operator() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:140 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x1095a)
#5 __invoke /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:110 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10809)
Previous write of size 8 at 0x7b4400001bc8 by main thread:
#0 free ??:? (completion_queue_test+0x4d65d4)
#1 ~AlarmImpl /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:44 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd551)
#2 grpc_impl::internal::AlarmImpl::Unref() /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:114 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd634)
#3 grpc_impl::internal::AlarmImpl::Destroy() /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:107 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd411)
#4 ~Alarm /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:156 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdcf62)
#5 ~Alarm /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:154 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdcfa9)
#6 std::default_delete<grpc_impl::Alarm>::operator()(grpc_impl::Alarm*) const /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/unique_ptr.h:81 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x29077)
#7 ~unique_ptr /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/unique_ptr.h:292 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27ef7)
#8 ~AsyncTimerFuture /proc/self/cwd/google/cloud/completion_queue.cc:38 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27430)
#9 void __gnu_cxx::new_allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>::destroy<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>(google::cloud::v1::(anonymous namespace)::AsyncTimerFuture*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/ext/new_allocator.h:153 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27ab1)
#10 void std::allocator_traits<std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture> >::destroy<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>(std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>&, google::cloud::v1::(anonymous namespace)::AsyncTimerFuture*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/alloc_traits.h:497 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27a49)
#11 std::_Sp_counted_ptr_inplace<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture, std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>, (__gnu_cxx::_Lock_policy)2>::_M_dispose() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:557 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x271b1)
#12 std::_Sp_counted_base<(__gnu_cxx::_Lock_policy)2>::_M_release() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:155 (completion_queue_test+0x581390)
#13 ~__shared_count /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:730 (completion_queue_test+0x581346)
#14 ~__shared_ptr /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:1169 (completion_queue_test+0x584a7d)
#15 google::cloud::v1::internal::CompletionQueueImpl::SimulateCompletion(bool) /proc/self/cwd/google/cloud/internal/completion_queue_impl.cc:127 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x335e7)
#16 google::cloud::v1::(anonymous namespace)::CompletionQueueTest_RunAsyncCompletionQueueDestroyed_Test::TestBody() /proc/self/cwd/google/cloud/completion_queue_test.cc:342 (completion_queue_test+0x575690)
#17 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe29d7)
#18 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xcffa2)
#19 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2508 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbf430)
#20 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2684 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbfd12)
#21 testing::TestSuite::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2816 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc0321)
#22 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:5338 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc9d31)
#23 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe6267)
#24 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xd2602)
#25 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4925 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc98d8)
#26 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2473 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xdfb)
#27 main /proc/self/cwd/external/com_google_googletest/googlemock/src/gmock_main.cc:63 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xda8)
Thread T3 'grpc_global_tim' (tid=4790, running) created by main thread at:
#0 pthread_create ??:? (completion_queue_test+0x4f3fa1)
#1 ThreadInternalsPosix /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:109 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10496)
#2 Thread /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:186 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10218)
#3 start_timer_thread_and_unlock() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:92 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf9417)
#4 start_threads() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:293 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf8fc5)
#5 grpc_timer_manager_init() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:311 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf8f4d)
#6 grpc_iomgr_start() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/iomgr.cc:69 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xdd243)
#7 grpc_init /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/surface/init.cc:162 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc.so+0x45f8)
#8 grpc::internal::GrpcLibrary::init() /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/grpc_library.h:34 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++.so+0x37fb3)
#9 GrpcLibraryCodegen /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/grpc_library.h:45 (completion_queue_test+0x583d9c)
#10 CompletionQueue /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/completion_queue_impl.h:247 (completion_queue_test+0x583c49)
#11 CompletionQueue /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/completion_queue_impl.h:108 (completion_queue_test+0x583bc7)
#12 CompletionQueueImpl /proc/self/cwd/./google/cloud/internal/completion_queue_impl.h:232 (completion_queue_test+0x583af4)
#13 MockCompletionQueue /proc/self/cwd/google/cloud/completion_queue_test.cc:30 (completion_queue_test+0x55cb0a)
#14 void __gnu_cxx::new_allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>::construct<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>(google::cloud::v1::(anonymous namespace)::MockCompletionQueue*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/ext/new_allocator.h:147 (completion_queue_test+0x55cac8)
#15 void std::allocator_traits<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> >::construct<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>(std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>&, google::cloud::v1::(anonymous namespace)::MockCompletionQueue*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/alloc_traits.h:484 (completion_queue_test+0x55c929)
#16 _Sp_counted_ptr_inplace<> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:548 (completion_queue_test+0x55c6f8)
#17 __shared_count<google::cloud::v1::(anonymous namespace)::MockCompletionQueue, std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:680 (completion_queue_test+0x55c5b2)
#18 __shared_ptr<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:1344 (completion_queue_test+0x55c553)
#19 shared_ptr<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:359 (completion_queue_test+0x55c500)
#20 std::shared_ptr<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> std::allocate_shared<google::cloud::v1::(anonymous namespace)::MockCompletionQueue, std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>>(std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> const&) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:701 (completion_queue_test+0x55c4c0)
#21 std::shared_ptr<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> std::make_shared<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:717 (completion_queue_test+0x55c38d)
#22 google::cloud::v1::(anonymous namespace)::CompletionQueueTest_RunAsyncCompletionQueueDestroyed_Test::TestBody() /proc/self/cwd/google/cloud/completion_queue_test.cc:332 (completion_queue_test+0x57562e)
#23 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe29d7)
#24 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xcffa2)
#25 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2508 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbf430)
#26 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2684 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbfd12)
#27 testing::TestSuite::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2816 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc0321)
#28 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:5338 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc9d31)
#29 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe6267)
#30 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xd2602)
#31 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4925 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc98d8)
#32 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2473 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xdfb)
#33 main /proc/self/cwd/external/com_google_googletest/googlemock/src/gmock_main.cc:63 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xda8)
SUMMARY: ThreadSanitizer: heap-use-after-free /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/exec_ctx.cc:150 in grpc_core::ExecCtx::Flush()
``` | 1.0 | heap-use-after-free in CompletionQueueTest.ShutdownWithReschedulingTimer - Are we holding the `grpc::Alarm` incorrectly?
[cloud-cpp/github/google-cloud-cpp/master/docker/tsan-presubmit](https://source.cloud.google.com/results/invocations/a5a82acb-c556-4c9a-9cdb-02ba5bd9fef5/targets/cloud-cpp%2Fgithub%2Fgoogle-cloud-cpp%2Fmaster%2Fdocker%2Ftsan-presubmit/log) failure:
```
[ RUN ] CompletionQueueTest.ShutdownWithReschedulingTimer
==================
WARNING: ThreadSanitizer: heap-use-after-free (pid=4743)
Read of size 8 at 0x7b4400001bc8 by thread T3:
#0 grpc_core::ExecCtx::Flush() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/exec_ctx.cc:150 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xd847a)
#1 run_some_timers() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:134 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf97d2)
#2 timer_main_loop() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:237 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf9593)
#3 timer_thread(void*) /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:284 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf94d7)
#4 operator() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:140 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x1095a)
#5 __invoke /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:110 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10809)
Previous write of size 8 at 0x7b4400001bc8 by main thread:
#0 free ??:? (completion_queue_test+0x4d65d4)
#1 ~AlarmImpl /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:44 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd551)
#2 grpc_impl::internal::AlarmImpl::Unref() /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:114 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd634)
#3 grpc_impl::internal::AlarmImpl::Destroy() /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:107 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdd411)
#4 ~Alarm /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:156 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdcf62)
#5 ~Alarm /proc/self/cwd/external/com_github_grpc_grpc/src/cpp/common/alarm.cc:154 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++_Ubase.so+0xdcfa9)
#6 std::default_delete<grpc_impl::Alarm>::operator()(grpc_impl::Alarm*) const /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/unique_ptr.h:81 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x29077)
#7 ~unique_ptr /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/unique_ptr.h:292 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27ef7)
#8 ~AsyncTimerFuture /proc/self/cwd/google/cloud/completion_queue.cc:38 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27430)
#9 void __gnu_cxx::new_allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>::destroy<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>(google::cloud::v1::(anonymous namespace)::AsyncTimerFuture*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/ext/new_allocator.h:153 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27ab1)
#10 void std::allocator_traits<std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture> >::destroy<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>(std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>&, google::cloud::v1::(anonymous namespace)::AsyncTimerFuture*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/alloc_traits.h:497 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x27a49)
#11 std::_Sp_counted_ptr_inplace<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture, std::allocator<google::cloud::v1::(anonymous namespace)::AsyncTimerFuture>, (__gnu_cxx::_Lock_policy)2>::_M_dispose() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:557 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x271b1)
#12 std::_Sp_counted_base<(__gnu_cxx::_Lock_policy)2>::_M_release() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:155 (completion_queue_test+0x581390)
#13 ~__shared_count /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:730 (completion_queue_test+0x581346)
#14 ~__shared_ptr /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:1169 (completion_queue_test+0x584a7d)
#15 google::cloud::v1::internal::CompletionQueueImpl::SimulateCompletion(bool) /proc/self/cwd/google/cloud/internal/completion_queue_impl.cc:127 (libgoogle_Scloud_Slibgoogle_Ucloud_Ucpp_Ugrpc_Uutils.so+0x335e7)
#16 google::cloud::v1::(anonymous namespace)::CompletionQueueTest_RunAsyncCompletionQueueDestroyed_Test::TestBody() /proc/self/cwd/google/cloud/completion_queue_test.cc:342 (completion_queue_test+0x575690)
#17 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe29d7)
#18 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xcffa2)
#19 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2508 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbf430)
#20 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2684 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbfd12)
#21 testing::TestSuite::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2816 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc0321)
#22 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:5338 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc9d31)
#23 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe6267)
#24 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xd2602)
#25 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4925 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc98d8)
#26 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2473 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xdfb)
#27 main /proc/self/cwd/external/com_google_googletest/googlemock/src/gmock_main.cc:63 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xda8)
Thread T3 'grpc_global_tim' (tid=4790, running) created by main thread at:
#0 pthread_create ??:? (completion_queue_test+0x4f3fa1)
#1 ThreadInternalsPosix /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:109 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10496)
#2 Thread /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/gprpp/thd_posix.cc:186 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgpr_Ubase.so+0x10218)
#3 start_timer_thread_and_unlock() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:92 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf9417)
#4 start_threads() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:293 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf8fc5)
#5 grpc_timer_manager_init() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/timer_manager.cc:311 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xf8f4d)
#6 grpc_iomgr_start() /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/iomgr.cc:69 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc_Ubase_Uc.so+0xdd243)
#7 grpc_init /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/surface/init.cc:162 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc.so+0x45f8)
#8 grpc::internal::GrpcLibrary::init() /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/grpc_library.h:34 (libexternal_Scom_Ugithub_Ugrpc_Ugrpc_Slibgrpc++.so+0x37fb3)
#9 GrpcLibraryCodegen /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/grpc_library.h:45 (completion_queue_test+0x583d9c)
#10 CompletionQueue /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/completion_queue_impl.h:247 (completion_queue_test+0x583c49)
#11 CompletionQueue /proc/self/cwd/external/com_github_grpc_grpc/include/grpcpp/impl/codegen/completion_queue_impl.h:108 (completion_queue_test+0x583bc7)
#12 CompletionQueueImpl /proc/self/cwd/./google/cloud/internal/completion_queue_impl.h:232 (completion_queue_test+0x583af4)
#13 MockCompletionQueue /proc/self/cwd/google/cloud/completion_queue_test.cc:30 (completion_queue_test+0x55cb0a)
#14 void __gnu_cxx::new_allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>::construct<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>(google::cloud::v1::(anonymous namespace)::MockCompletionQueue*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/ext/new_allocator.h:147 (completion_queue_test+0x55cac8)
#15 void std::allocator_traits<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> >::construct<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>(std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>&, google::cloud::v1::(anonymous namespace)::MockCompletionQueue*) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/alloc_traits.h:484 (completion_queue_test+0x55c929)
#16 _Sp_counted_ptr_inplace<> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:548 (completion_queue_test+0x55c6f8)
#17 __shared_count<google::cloud::v1::(anonymous namespace)::MockCompletionQueue, std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:680 (completion_queue_test+0x55c5b2)
#18 __shared_ptr<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr_base.h:1344 (completion_queue_test+0x55c553)
#19 shared_ptr<std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>> /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:359 (completion_queue_test+0x55c500)
#20 std::shared_ptr<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> std::allocate_shared<google::cloud::v1::(anonymous namespace)::MockCompletionQueue, std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>>(std::allocator<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> const&) /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:701 (completion_queue_test+0x55c4c0)
#21 std::shared_ptr<google::cloud::v1::(anonymous namespace)::MockCompletionQueue> std::make_shared<google::cloud::v1::(anonymous namespace)::MockCompletionQueue>() /usr/bin/../lib/gcc/x86_64-redhat-linux/9/../../../../include/c++/9/bits/shared_ptr.h:717 (completion_queue_test+0x55c38d)
#22 google::cloud::v1::(anonymous namespace)::CompletionQueueTest_RunAsyncCompletionQueueDestroyed_Test::TestBody() /proc/self/cwd/google/cloud/completion_queue_test.cc:332 (completion_queue_test+0x57562e)
#23 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe29d7)
#24 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xcffa2)
#25 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2508 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbf430)
#26 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2684 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xbfd12)
#27 testing::TestSuite::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2816 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc0321)
#28 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:5338 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc9d31)
#29 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2433 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xe6267)
#30 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2469 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xd2602)
#31 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4925 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest.so+0xc98d8)
#32 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2473 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xdfb)
#33 main /proc/self/cwd/external/com_google_googletest/googlemock/src/gmock_main.cc:63 (libexternal_Scom_Ugoogle_Ugoogletest_Slibgtest_Umain.so+0xda8)
SUMMARY: ThreadSanitizer: heap-use-after-free /proc/self/cwd/external/com_github_grpc_grpc/src/core/lib/iomgr/exec_ctx.cc:150 in grpc_core::ExecCtx::Flush()
``` | non_defect | heap use after free in completionqueuetest shutdownwithreschedulingtimer are we holding the grpc alarm incorrectly failure completionqueuetest shutdownwithreschedulingtimer warning threadsanitizer heap use after free pid read of size at by thread grpc core execctx flush proc self cwd external com github grpc grpc src core lib iomgr exec ctx cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so run some timers proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so timer main loop proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so timer thread void proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so operator proc self cwd external com github grpc grpc src core lib gprpp thd posix cc libexternal scom ugithub ugrpc ugrpc slibgpr ubase so invoke proc self cwd external com github grpc grpc src core lib gprpp thd posix cc libexternal scom ugithub ugrpc ugrpc slibgpr ubase so previous write of size at by main thread free completion queue test alarmimpl proc self cwd external com github grpc grpc src cpp common alarm cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase so grpc impl internal alarmimpl unref proc self cwd external com github grpc grpc src cpp common alarm cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase so grpc impl internal alarmimpl destroy proc self cwd external com github grpc grpc src cpp common alarm cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase so alarm proc self cwd external com github grpc grpc src cpp common alarm cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase so alarm proc self cwd external com github grpc grpc src cpp common alarm cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase so std default delete operator grpc impl alarm const usr bin lib gcc redhat linux include c bits unique ptr h libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so unique ptr usr bin lib gcc redhat linux include c bits unique ptr h libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so asynctimerfuture proc self cwd google cloud completion queue cc libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so void gnu cxx new allocator destroy google cloud anonymous namespace asynctimerfuture usr bin lib gcc redhat linux include c ext new allocator h libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so void std allocator traits destroy std allocator google cloud anonymous namespace asynctimerfuture usr bin lib gcc redhat linux include c bits alloc traits h libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so std sp counted ptr inplace gnu cxx lock policy m dispose usr bin lib gcc redhat linux include c bits shared ptr base h libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so std sp counted base m release usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test shared count usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test shared ptr usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test google cloud internal completionqueueimpl simulatecompletion bool proc self cwd google cloud internal completion queue impl cc libgoogle scloud slibgoogle ucloud ucpp ugrpc uutils so google cloud anonymous namespace completionqueuetest runasynccompletionqueuedestroyed test testbody proc self cwd google cloud completion queue test cc completion queue test void testing internal handlesehexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so void testing internal handleexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing test run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing testinfo run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing testsuite run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing internal unittestimpl runalltests proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so bool testing internal handlesehexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing unittest run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so run all tests proc self cwd external com google googletest googletest include gtest gtest h libexternal scom ugoogle ugoogletest slibgtest umain so main proc self cwd external com google googletest googlemock src gmock main cc libexternal scom ugoogle ugoogletest slibgtest umain so thread grpc global tim tid running created by main thread at pthread create completion queue test threadinternalsposix proc self cwd external com github grpc grpc src core lib gprpp thd posix cc libexternal scom ugithub ugrpc ugrpc slibgpr ubase so thread proc self cwd external com github grpc grpc src core lib gprpp thd posix cc libexternal scom ugithub ugrpc ugrpc slibgpr ubase so start timer thread and unlock proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so start threads proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so grpc timer manager init proc self cwd external com github grpc grpc src core lib iomgr timer manager cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so grpc iomgr start proc self cwd external com github grpc grpc src core lib iomgr iomgr cc libexternal scom ugithub ugrpc ugrpc slibgrpc ubase uc so grpc init proc self cwd external com github grpc grpc src core lib surface init cc libexternal scom ugithub ugrpc ugrpc slibgrpc so grpc internal grpclibrary init proc self cwd external com github grpc grpc include grpcpp impl grpc library h libexternal scom ugithub ugrpc ugrpc slibgrpc so grpclibrarycodegen proc self cwd external com github grpc grpc include grpcpp impl codegen grpc library h completion queue test completionqueue proc self cwd external com github grpc grpc include grpcpp impl codegen completion queue impl h completion queue test completionqueue proc self cwd external com github grpc grpc include grpcpp impl codegen completion queue impl h completion queue test completionqueueimpl proc self cwd google cloud internal completion queue impl h completion queue test mockcompletionqueue proc self cwd google cloud completion queue test cc completion queue test void gnu cxx new allocator construct google cloud anonymous namespace mockcompletionqueue usr bin lib gcc redhat linux include c ext new allocator h completion queue test void std allocator traits construct std allocator google cloud anonymous namespace mockcompletionqueue usr bin lib gcc redhat linux include c bits alloc traits h completion queue test sp counted ptr inplace usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test shared count usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test shared ptr usr bin lib gcc redhat linux include c bits shared ptr base h completion queue test shared ptr usr bin lib gcc redhat linux include c bits shared ptr h completion queue test std shared ptr std allocate shared std allocator const usr bin lib gcc redhat linux include c bits shared ptr h completion queue test std shared ptr std make shared usr bin lib gcc redhat linux include c bits shared ptr h completion queue test google cloud anonymous namespace completionqueuetest runasynccompletionqueuedestroyed test testbody proc self cwd google cloud completion queue test cc completion queue test void testing internal handlesehexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so void testing internal handleexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing test run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing testinfo run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing testsuite run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing internal unittestimpl runalltests proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so bool testing internal handlesehexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so testing unittest run proc self cwd external com google googletest googletest src gtest cc libexternal scom ugoogle ugoogletest slibgtest so run all tests proc self cwd external com google googletest googletest include gtest gtest h libexternal scom ugoogle ugoogletest slibgtest umain so main proc self cwd external com google googletest googlemock src gmock main cc libexternal scom ugoogle ugoogletest slibgtest umain so summary threadsanitizer heap use after free proc self cwd external com github grpc grpc src core lib iomgr exec ctx cc in grpc core execctx flush | 0 |
140,498 | 18,902,265,613 | IssuesEvent | 2021-11-16 03:17:45 | jinhogate/pizza_angular | https://api.github.com/repos/jinhogate/pizza_angular | opened | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz | security vulnerability | ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: pizza_angular/package.json</p>
<p>Path to vulnerable library: pizza_angular/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.6.8.tgz (Root Library)
- less-3.8.1.tgz
- request-2.88.0.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: pizza_angular/package.json</p>
<p>Path to vulnerable library: pizza_angular/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.6.8.tgz (Root Library)
- less-3.8.1.tgz
- request-2.88.0.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file pizza angular package json path to vulnerable library pizza angular node modules json schema package json dependency hierarchy build angular tgz root library less tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
37,879 | 8,559,746,799 | IssuesEvent | 2018-11-08 22:15:57 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | capped_distance array: kernel exploding for a specific frame in this trajectory | Component-lib defect | **Expected behavior**
capped_distance should return something for each step of the trajectory
**Actual behavior**
when iterating over the trajectory, the kernel dies as the 8th frame is being processed - so initially I suspected a memory issue.
however, when running only the 8th frame, the same thing (see below) happens.
I have used other analysis functions on this trajectory including distance_array, so it seems specific to the recent implementation of capped_distance.
**Code to reproduce the behavior**
``` python
import MDAnalysis as mda
from MDAnalysis.lib import distances
u = mda.Universe(TOP, TRJ)
u.trajectory[7]
A=u.atoms.select_atoms("type os oh")
H=u.atoms.select_atoms("type HW")
#this works
distances.distance_array(H.positions, A.positions, box=u.dimensions)
#this doesn't
capped_distance(H.positions, A.positions, max_cutoff=3.0, box=u.dimensions)
```
**Currently version of MDAnalysis**
- 0.19.0
- Python 3.6.6 :: Anaconda custom (64-bit)
- Mac OS X Mojave
[TOP](https://github.com/MDAnalysis/mdanalysis/files/2548915/PEG.prmtop.zip)
[TRJ](https://github.com/MDAnalysis/mdanalysis/files/2548810/PEG_03_prod.nc.zip)
| 1.0 | capped_distance array: kernel exploding for a specific frame in this trajectory - **Expected behavior**
capped_distance should return something for each step of the trajectory
**Actual behavior**
when iterating over the trajectory, the kernel dies as the 8th frame is being processed - so initially I suspected a memory issue.
however, when running only the 8th frame, the same thing (see below) happens.
I have used other analysis functions on this trajectory including distance_array, so it seems specific to the recent implementation of capped_distance.
**Code to reproduce the behavior**
``` python
import MDAnalysis as mda
from MDAnalysis.lib import distances
u = mda.Universe(TOP, TRJ)
u.trajectory[7]
A=u.atoms.select_atoms("type os oh")
H=u.atoms.select_atoms("type HW")
#this works
distances.distance_array(H.positions, A.positions, box=u.dimensions)
#this doesn't
capped_distance(H.positions, A.positions, max_cutoff=3.0, box=u.dimensions)
```
**Currently version of MDAnalysis**
- 0.19.0
- Python 3.6.6 :: Anaconda custom (64-bit)
- Mac OS X Mojave
[TOP](https://github.com/MDAnalysis/mdanalysis/files/2548915/PEG.prmtop.zip)
[TRJ](https://github.com/MDAnalysis/mdanalysis/files/2548810/PEG_03_prod.nc.zip)
| defect | capped distance array kernel exploding for a specific frame in this trajectory expected behavior capped distance should return something for each step of the trajectory actual behavior when iterating over the trajectory the kernel dies as the frame is being processed so initially i suspected a memory issue however when running only the frame the same thing see below happens i have used other analysis functions on this trajectory including distance array so it seems specific to the recent implementation of capped distance code to reproduce the behavior python import mdanalysis as mda from mdanalysis lib import distances u mda universe top trj u trajectory a u atoms select atoms type os oh h u atoms select atoms type hw this works distances distance array h positions a positions box u dimensions this doesn t capped distance h positions a positions max cutoff box u dimensions currently version of mdanalysis python anaconda custom bit mac os x mojave | 1 |
27,711 | 5,078,795,333 | IssuesEvent | 2016-12-28 16:52:18 | opendatakit/opendatakit | https://api.github.com/repos/opendatakit/opendatakit | closed | date format in prompt view doesn't follow system defaults | 1.2 Collect Priority-Medium Type-Defect | *Migrated to opendatakit/collect#320 by [spacetelescope/github-issues-import](https://github.com/spacetelescope/github-issues-import)*
Originally reported on Google Code with ID 638
```
What steps will reproduce the problem?
1.enter a date
2.view prompts (go to prompts)
What is the expected output? What do you see instead?
My system is set to display dates in YYYY-MM-DD, so I expect to view the same. Right
now, dates are DD-MM-YYYY.
What version of the product are you using? On what operating system?
Collect 1.2 (1010), Android 2.3.7
Please provide any additional information below.
```
Reported by `Nikolai.Go` on 2012-07-16 00:02:04
| 1.0 | date format in prompt view doesn't follow system defaults - *Migrated to opendatakit/collect#320 by [spacetelescope/github-issues-import](https://github.com/spacetelescope/github-issues-import)*
Originally reported on Google Code with ID 638
```
What steps will reproduce the problem?
1.enter a date
2.view prompts (go to prompts)
What is the expected output? What do you see instead?
My system is set to display dates in YYYY-MM-DD, so I expect to view the same. Right
now, dates are DD-MM-YYYY.
What version of the product are you using? On what operating system?
Collect 1.2 (1010), Android 2.3.7
Please provide any additional information below.
```
Reported by `Nikolai.Go` on 2012-07-16 00:02:04
| defect | date format in prompt view doesn t follow system defaults migrated to opendatakit collect by originally reported on google code with id what steps will reproduce the problem enter a date view prompts go to prompts what is the expected output what do you see instead my system is set to display dates in yyyy mm dd so i expect to view the same right now dates are dd mm yyyy what version of the product are you using on what operating system collect android please provide any additional information below reported by nikolai go on | 1 |
35,186 | 7,642,306,179 | IssuesEvent | 2018-05-08 08:49:51 | primefaces/primereact | https://api.github.com/repos/primefaces/primereact | closed | TabView activeIndex setting programmatically will not work the second time after manually choosing another tab | defect | **I'm submitting a ...** (check one with "x")
```
[x ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
https://plnkr.co/edit/C6qjkdLtbmy92Hogdnjb?p=preview
**Current behavior**
Once you set activeIndex programmatically and then switch tabs by clicking, you can't set the index to the same value you set previously because the newly received props is not compared to the state but to the original props, which hasn't changed since you changed the index from within the component.
**Expected behavior**
Switching tabs programmatically works more than once even after switching them manually.
**Minimal reproduction of the problem with instructions**
1. Create at least 2 tab tabView.
2. Set any tab programmatically
3. Select another tab
4. Try selecting the same tab from p2. programmatically.
| 1.0 | TabView activeIndex setting programmatically will not work the second time after manually choosing another tab - **I'm submitting a ...** (check one with "x")
```
[x ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
https://plnkr.co/edit/C6qjkdLtbmy92Hogdnjb?p=preview
**Current behavior**
Once you set activeIndex programmatically and then switch tabs by clicking, you can't set the index to the same value you set previously because the newly received props is not compared to the state but to the original props, which hasn't changed since you changed the index from within the component.
**Expected behavior**
Switching tabs programmatically works more than once even after switching them manually.
**Minimal reproduction of the problem with instructions**
1. Create at least 2 tab tabView.
2. Set any tab programmatically
3. Select another tab
4. Try selecting the same tab from p2. programmatically.
| defect | tabview activeindex setting programmatically will not work the second time after manually choosing another tab i m submitting a check one with x bug report feature request support request please do not submit support request here instead see current behavior once you set activeindex programmatically and then switch tabs by clicking you can t set the index to the same value you set previously because the newly received props is not compared to the state but to the original props which hasn t changed since you changed the index from within the component expected behavior switching tabs programmatically works more than once even after switching them manually minimal reproduction of the problem with instructions create at least tab tabview set any tab programmatically select another tab try selecting the same tab from programmatically | 1 |
245,876 | 18,796,940,732 | IssuesEvent | 2021-11-08 23:55:09 | pokt-network/pocket-core-func-tests | https://api.github.com/repos/pokt-network/pocket-core-func-tests | closed | Pocket Core RC-0.6.0 Functional Tests Scope | documentation enhancement | ### Documentation For Pocket Core RC-0.6.0 Functional Tests Scope
### Leader
[Emanuel Medrano: Support Engineering Specialist]
### Date(s)
02/15/2021
### Participants
- [Emanuel Medrano: Support Engineering Specialist]
## Survey
### Category
- [ ] Infrastructure
- [x] Protocol/Blockchain Dev
- [ ] App Solutions
### LEVEL
- [x] RC
- [ ] STABLE | 1.0 | Pocket Core RC-0.6.0 Functional Tests Scope - ### Documentation For Pocket Core RC-0.6.0 Functional Tests Scope
### Leader
[Emanuel Medrano: Support Engineering Specialist]
### Date(s)
02/15/2021
### Participants
- [Emanuel Medrano: Support Engineering Specialist]
## Survey
### Category
- [ ] Infrastructure
- [x] Protocol/Blockchain Dev
- [ ] App Solutions
### LEVEL
- [x] RC
- [ ] STABLE | non_defect | pocket core rc functional tests scope documentation for pocket core rc functional tests scope leader date s participants survey category infrastructure protocol blockchain dev app solutions level rc stable | 0 |
319,551 | 23,778,556,055 | IssuesEvent | 2022-09-02 00:22:12 | wp-graphql/wp-graphql | https://api.github.com/repos/wp-graphql/wp-graphql | closed | WPML and action_monitor fetching | Documentation Needs Discussion WPML stale | WPML redirects when WPGraphQL fetches action_monitor posts
Adding the snippet below fixes the issue.
```
add_filter( 'wpml_is_redirected', function( $is_redirect ) {
if ( is_graphql_request() ) {
return false;
}
return $is_redirect;
});
``` | 1.0 | WPML and action_monitor fetching - WPML redirects when WPGraphQL fetches action_monitor posts
Adding the snippet below fixes the issue.
```
add_filter( 'wpml_is_redirected', function( $is_redirect ) {
if ( is_graphql_request() ) {
return false;
}
return $is_redirect;
});
``` | non_defect | wpml and action monitor fetching wpml redirects when wpgraphql fetches action monitor posts adding the snippet below fixes the issue add filter wpml is redirected function is redirect if is graphql request return false return is redirect | 0 |
56,303 | 15,020,011,886 | IssuesEvent | 2021-02-01 14:14:07 | mozilla-lockwise/lockwise-ios | https://api.github.com/repos/mozilla-lockwise/lockwise-ios | reopened | Item Detail view reloads when returning to app | archived defect | Steps to reproduce:
- open an item detail
- tap on the website name to open the website
- return to Lockwise, either via the app switcher or the tiny "Back" option in the status bar
**expected**:
no visual change in Lockwise display
**actual**:
the display flickers when returning | 1.0 | Item Detail view reloads when returning to app - Steps to reproduce:
- open an item detail
- tap on the website name to open the website
- return to Lockwise, either via the app switcher or the tiny "Back" option in the status bar
**expected**:
no visual change in Lockwise display
**actual**:
the display flickers when returning | defect | item detail view reloads when returning to app steps to reproduce open an item detail tap on the website name to open the website return to lockwise either via the app switcher or the tiny back option in the status bar expected no visual change in lockwise display actual the display flickers when returning | 1 |
9,495 | 2,615,153,576 | IssuesEvent | 2015-03-01 06:31:21 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | sending eapol request - recive timeout ocurred | auto-migrated Priority-Triage Type-Defect | ```
Using reaver-1.04 on Bactrack 4 (R1 y R2) with suported driver (rtl8187) with
very god signal -33 to my router OBSERVA TECOM AW4062. Starts ok after SENDING
EAPOL REQUEST - WARNING: Receive timeout ocurred
SENDING EAPOL REQUEST - WARNING: Receive timeout ocurred
SENDING EAPOL REQUEST - WARNING: Receive timeout ocurred
After 25 failures : WPS Transaction failec (code 0x2)
```
Original issue reported on code.google.com by `faustinp...@gmail.com` on 10 Feb 2012 at 5:06 | 1.0 | sending eapol request - recive timeout ocurred - ```
Using reaver-1.04 on Bactrack 4 (R1 y R2) with suported driver (rtl8187) with
very god signal -33 to my router OBSERVA TECOM AW4062. Starts ok after SENDING
EAPOL REQUEST - WARNING: Receive timeout ocurred
SENDING EAPOL REQUEST - WARNING: Receive timeout ocurred
SENDING EAPOL REQUEST - WARNING: Receive timeout ocurred
After 25 failures : WPS Transaction failec (code 0x2)
```
Original issue reported on code.google.com by `faustinp...@gmail.com` on 10 Feb 2012 at 5:06 | defect | sending eapol request recive timeout ocurred using reaver on bactrack y with suported driver with very god signal to my router observa tecom starts ok after sending eapol request warning receive timeout ocurred sending eapol request warning receive timeout ocurred sending eapol request warning receive timeout ocurred after failures wps transaction failec code original issue reported on code google com by faustinp gmail com on feb at | 1 |
45,262 | 12,691,437,383 | IssuesEvent | 2020-06-21 17:02:03 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | 4.0: $this->disableErrorHandlerMiddleware(); has no effect in IntegrationTestCase | defect | `$this->disableErrorHandlerMiddleware();` has no effect in 4.x right now.
version: latest master.
```php
/**
* @var array
*/
protected $fixtures = [
//'plugin.Data.Countries',
];
```
Commenting it out causes:
> message: "Cannot describe countries. It has 0 columns."
code: 500
But instead of failing hard, it gets caught be middleware, and all I get is
> Failed asserting that `200` matches response status code `500`.
instead of the expected exception with details on why this 500 is caused. Hiding the message I saw through debugger makes it hard to find actual issues. | 1.0 | 4.0: $this->disableErrorHandlerMiddleware(); has no effect in IntegrationTestCase - `$this->disableErrorHandlerMiddleware();` has no effect in 4.x right now.
version: latest master.
```php
/**
* @var array
*/
protected $fixtures = [
//'plugin.Data.Countries',
];
```
Commenting it out causes:
> message: "Cannot describe countries. It has 0 columns."
code: 500
But instead of failing hard, it gets caught be middleware, and all I get is
> Failed asserting that `200` matches response status code `500`.
instead of the expected exception with details on why this 500 is caused. Hiding the message I saw through debugger makes it hard to find actual issues. | defect | this disableerrorhandlermiddleware has no effect in integrationtestcase this disableerrorhandlermiddleware has no effect in x right now version latest master php var array protected fixtures plugin data countries commenting it out causes message cannot describe countries it has columns code but instead of failing hard it gets caught be middleware and all i get is failed asserting that matches response status code instead of the expected exception with details on why this is caused hiding the message i saw through debugger makes it hard to find actual issues | 1 |
45,058 | 7,156,867,432 | IssuesEvent | 2018-01-26 17:45:42 | cerner/clara-rules | https://api.github.com/repos/cerner/clara-rules | closed | Update copyright year? | documentation | Just curious whether the copyright year (2016) can be updated to 2018. :)
Shows 2016 here: https://github.com/cerner/clara-rules/blob/master/README.md. Guessing other places as well. | 1.0 | Update copyright year? - Just curious whether the copyright year (2016) can be updated to 2018. :)
Shows 2016 here: https://github.com/cerner/clara-rules/blob/master/README.md. Guessing other places as well. | non_defect | update copyright year just curious whether the copyright year can be updated to shows here guessing other places as well | 0 |
10,014 | 13,043,884,184 | IssuesEvent | 2020-07-29 02:56:51 | tikv/tikv | https://api.github.com/repos/tikv/tikv | closed | UCP: Migrate scalar function `WeekWithMode` from TiDB | challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor |
## Description
Port the scalar function `WeekWithMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| 2.0 | UCP: Migrate scalar function `WeekWithMode` from TiDB -
## Description
Port the scalar function `WeekWithMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| non_defect | ucp migrate scalar function weekwithmode from tidb description port the scalar function weekwithmode from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb | 0 |
2,669 | 2,607,935,290 | IssuesEvent | 2015-02-26 00:28:34 | chrsmithdemos/minify | https://api.github.com/repos/chrsmithdemos/minify | closed | Safari reports 304 responses as missing Content-Type | auto-migrated Priority-Medium Type-Defect | ```
Minify version: '2.1.4';
PHP version: 5.2.12
What steps will reproduce the problem?
1. Start Safari 5.0.2
2. Empty cache or disable cache
3. Show webinfo window (developer tools)
4. Open source files and enable profiling
5. Open http://crashplan.probackup.nl/
Expected output:
No warnings
Actual output:
Warning: Source file interpreted as stylesheet, but transferred with MIME type
text/html [g=css&0]
Did any unit tests FAIL? (Please do not post the full list)
No
```
-----
Original issue reported on code.google.com by `i...@crashplan.nl` on 17 Nov 2010 at 9:07 | 1.0 | Safari reports 304 responses as missing Content-Type - ```
Minify version: '2.1.4';
PHP version: 5.2.12
What steps will reproduce the problem?
1. Start Safari 5.0.2
2. Empty cache or disable cache
3. Show webinfo window (developer tools)
4. Open source files and enable profiling
5. Open http://crashplan.probackup.nl/
Expected output:
No warnings
Actual output:
Warning: Source file interpreted as stylesheet, but transferred with MIME type
text/html [g=css&0]
Did any unit tests FAIL? (Please do not post the full list)
No
```
-----
Original issue reported on code.google.com by `i...@crashplan.nl` on 17 Nov 2010 at 9:07 | defect | safari reports responses as missing content type minify version php version what steps will reproduce the problem start safari empty cache or disable cache show webinfo window developer tools open source files and enable profiling open expected output no warnings actual output warning source file interpreted as stylesheet but transferred with mime type text html did any unit tests fail please do not post the full list no original issue reported on code google com by i crashplan nl on nov at | 1 |
53,373 | 22,768,816,299 | IssuesEvent | 2022-07-08 08:01:52 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | `az ams account show` | Media Services customer-reported needs-author-feedback no-recent-activity CXP Attention Auto-Assign | ### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az ams account show`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
'MediaservicesOperations' object has no attribute 'get_by_subscription'
Traceback (most recent call last):
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\knack/cli.py", line 231, in invoke
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 663, in execute
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 718, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/ams/_exception_handler.py", line 16, in ams_exception_handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 697, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 333, in __call__
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 363, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/arm.py", line 429, in show_exception_handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 361, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/ams/operations/account.py", line 17, in get_mediaservice
AttributeError: 'MediaservicesOperations' object has no attribute 'get_by_subscription'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az ams account show --name {}`
## Expected Behavior
List all ams account
## Environment Summary
```
Windows-10-10.0.19044-SP0
Python 3.10.4
Installer:
azure-cli 2.37.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
| 1.0 | `az ams account show` - ### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az ams account show`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
'MediaservicesOperations' object has no attribute 'get_by_subscription'
Traceback (most recent call last):
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\knack/cli.py", line 231, in invoke
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 663, in execute
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 718, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/ams/_exception_handler.py", line 16, in ams_exception_handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 697, in _run_job
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/__init__.py", line 333, in __call__
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 363, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/arm.py", line 429, in show_exception_handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/core/commands/command_operation.py", line 361, in handler
File "D:\a\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/cli/command_modules/ams/operations/account.py", line 17, in get_mediaservice
AttributeError: 'MediaservicesOperations' object has no attribute 'get_by_subscription'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az ams account show --name {}`
## Expected Behavior
List all ams account
## Environment Summary
```
Windows-10-10.0.19044-SP0
Python 3.10.4
Installer:
azure-cli 2.37.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
| non_defect | az ams account show this is autogenerated please review and update as needed describe the bug command name az ams account show errors the command failed with an unexpected error here is the traceback mediaservicesoperations object has no attribute get by subscription traceback most recent call last file d a s build scripts windows artifacts cli lib site packages knack cli py line in invoke file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in execute file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in run jobs serially file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in run job file d a s build scripts windows artifacts cli lib site packages azure cli command modules ams exception handler py line in ams exception handler file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in run job file d a s build scripts windows artifacts cli lib site packages azure cli core commands init py line in call file d a s build scripts windows artifacts cli lib site packages azure cli core commands command operation py line in handler file d a s build scripts windows artifacts cli lib site packages azure cli core commands arm py line in show exception handler file d a s build scripts windows artifacts cli lib site packages azure cli core commands command operation py line in handler file d a s build scripts windows artifacts cli lib site packages azure cli command modules ams operations account py line in get mediaservice attributeerror mediaservicesoperations object has no attribute get by subscription to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az ams account show name expected behavior list all ams account environment summary windows python installer azure cli additional context | 0 |
196,894 | 14,895,625,105 | IssuesEvent | 2021-01-21 09:21:20 | WeAreSnook/connected-camden | https://api.github.com/repos/WeAreSnook/connected-camden | closed | [21] Test printing on a range of devices and printers | testing | > As a system I need to print correctly regardless of what printer / device people use so people can use the printed routes
Whilst we don't have access to a printer or a way to test that route printing works we need to manually test route printing on a range of devices.
## Acceptance criteria
- [x] We have tested printing on at least 3 different printers
- [x] We have tested printing in black and white
- [x] We have tested printing in colour
- [x] We have tested printing from a Windows computer
- [x] We have tested printing from a Mac computer
- [x] We have tested printing a long route
- [x] We have tested printing a short route
## Notes and links
- We need to find people who have printers. We should get the Camden team involved in testing printing. | 1.0 | [21] Test printing on a range of devices and printers - > As a system I need to print correctly regardless of what printer / device people use so people can use the printed routes
Whilst we don't have access to a printer or a way to test that route printing works we need to manually test route printing on a range of devices.
## Acceptance criteria
- [x] We have tested printing on at least 3 different printers
- [x] We have tested printing in black and white
- [x] We have tested printing in colour
- [x] We have tested printing from a Windows computer
- [x] We have tested printing from a Mac computer
- [x] We have tested printing a long route
- [x] We have tested printing a short route
## Notes and links
- We need to find people who have printers. We should get the Camden team involved in testing printing. | non_defect | test printing on a range of devices and printers as a system i need to print correctly regardless of what printer device people use so people can use the printed routes whilst we don t have access to a printer or a way to test that route printing works we need to manually test route printing on a range of devices acceptance criteria we have tested printing on at least different printers we have tested printing in black and white we have tested printing in colour we have tested printing from a windows computer we have tested printing from a mac computer we have tested printing a long route we have tested printing a short route notes and links we need to find people who have printers we should get the camden team involved in testing printing | 0 |
68,312 | 8,248,619,492 | IssuesEvent | 2018-09-11 18:59:22 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | closed | PD Analytics: Fullstory Opt-In | feature large protocol designer | As a protocol designer, I would like to be able to opt-in to share my usage data with Opentrons via Fullstory.
## Acceptance Criteria
- Add FullStory tracking to Protocol Designer
- Users prompted to opt in to share session data with Opentrons the first time they open the web app
- User able to turn session tracking on or off in the app
-- Add a new "Settings" tab to the bottom of our top level nav. The privacy list item will open a privacy card in the page.
## Implementation Notes
- [Link to code snippet and guide](https://app.fullstory.com/ui/B4RHZ/settings/general)
- [Discussion of FullStory and SPA](https://help.fullstory.com/using/does-fullstory-work-with-my-complex-site) (actually seems like a non-issue):
- Get account info from Alfie
## Design
Initial card upon opening session: https://zpl.io/aX4wR0X
I thought it made more sense as a card. We use modals when UI needs to go on top of something. If you don't have a protocol open yet there's no need for a modal.
Settings tab: https://zpl.io/25rGl4o | 1.0 | PD Analytics: Fullstory Opt-In - As a protocol designer, I would like to be able to opt-in to share my usage data with Opentrons via Fullstory.
## Acceptance Criteria
- Add FullStory tracking to Protocol Designer
- Users prompted to opt in to share session data with Opentrons the first time they open the web app
- User able to turn session tracking on or off in the app
-- Add a new "Settings" tab to the bottom of our top level nav. The privacy list item will open a privacy card in the page.
## Implementation Notes
- [Link to code snippet and guide](https://app.fullstory.com/ui/B4RHZ/settings/general)
- [Discussion of FullStory and SPA](https://help.fullstory.com/using/does-fullstory-work-with-my-complex-site) (actually seems like a non-issue):
- Get account info from Alfie
## Design
Initial card upon opening session: https://zpl.io/aX4wR0X
I thought it made more sense as a card. We use modals when UI needs to go on top of something. If you don't have a protocol open yet there's no need for a modal.
Settings tab: https://zpl.io/25rGl4o | non_defect | pd analytics fullstory opt in as a protocol designer i would like to be able to opt in to share my usage data with opentrons via fullstory acceptance criteria add fullstory tracking to protocol designer users prompted to opt in to share session data with opentrons the first time they open the web app user able to turn session tracking on or off in the app add a new settings tab to the bottom of our top level nav the privacy list item will open a privacy card in the page implementation notes actually seems like a non issue get account info from alfie design initial card upon opening session i thought it made more sense as a card we use modals when ui needs to go on top of something if you don t have a protocol open yet there s no need for a modal settings tab | 0 |
6,174 | 2,610,222,707 | IssuesEvent | 2015-02-26 19:10:38 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | гурмания 2 полная версия | auto-migrated Priority-Medium Type-Defect | ```
'''Геодар Егоров'''
День добрый никак не могу найти .гурмания 2
полная версия. как то выкладывали уже
'''Альберт Комаров'''
Качай тут http://bit.ly/1aVKsBb
'''воин Ковалёв'''
Спасибо вроде то но просит телефон вводить
'''Аверкий Яковлев'''
Неа все ок у меня ничего не списало
'''Альфред Орехов'''
Не это не влияет на баланс
Информация о файле: гурмания 2 полная версия
Загружен: В этом месяце
Скачан раз: 1046
Рейтинг: 526
Средняя скорость скачивания: 1380
Похожих файлов: 29
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 4:09 | 1.0 | гурмания 2 полная версия - ```
'''Геодар Егоров'''
День добрый никак не могу найти .гурмания 2
полная версия. как то выкладывали уже
'''Альберт Комаров'''
Качай тут http://bit.ly/1aVKsBb
'''воин Ковалёв'''
Спасибо вроде то но просит телефон вводить
'''Аверкий Яковлев'''
Неа все ок у меня ничего не списало
'''Альфред Орехов'''
Не это не влияет на баланс
Информация о файле: гурмания 2 полная версия
Загружен: В этом месяце
Скачан раз: 1046
Рейтинг: 526
Средняя скорость скачивания: 1380
Похожих файлов: 29
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 4:09 | defect | гурмания полная версия геодар егоров день добрый никак не могу найти гурмания полная версия как то выкладывали уже альберт комаров качай тут воин ковалёв спасибо вроде то но просит телефон вводить аверкий яковлев неа все ок у меня ничего не списало альфред орехов не это не влияет на баланс информация о файле гурмания полная версия загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
44,519 | 12,223,146,245 | IssuesEvent | 2020-05-02 16:19:17 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | ValueError 'k exceeds matrix dimensions' for sparse.diagonal() when 0 in sparse.shape | defect good first issue scipy.sparse | When a sparse matrix has a 0 in its shape, such as `(0, 0)`, `(0, 1)` or `(1, 0)`, calling `diagonal()` fails. This differs to `np.diag` on the equivalent dense array, which succeeds, returning an empty array.
The best behaviour here seems like it'd be open for debate, but it's unfortunate that the default `diagonal()` method doesn't work on every square sparse matrix. It can require adding special cases/conditionals around `.diagonal()` calls, such as https://github.com/stellargraph/stellargraph/pull/1378.
#### Reproducing code example:
Minimal:
```python
import scipy.sparse as sps
import numpy as np
m = sps.csr_matrix((0, 0))
print(np.diag(m.todense()).shape) # (0,)
m.diagonal() # ValueError: k exceeds matrix dimensions
```
"Complete" tests:
```python
import scipy.sparse as sps
# check all the sparse matrix classes
classes = [
sps.bsr_matrix,
sps.coo_matrix,
sps.csc_matrix,
sps.csr_matrix,
sps.dia_matrix,
sps.dok_matrix,
sps.lil_matrix,
]
for cls in classes:
try:
cls((0,0)).diagonal()
except Exception as e:
msg = e
else:
msg = None
print(f"{cls.__name__}: {msg!r}")
# For completeness, non-(0, 0) empty matrices
m = sps.csr_matrix((1, 0))
m.diagonal() # ValueError: k exceeds matrix dimensions
m = sps.csr_matrix((0, 1))
m.diagonal() # ValueError: k exceeds matrix dimensions
```
#### Error message:
Minimal example:
```
(0,)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-44-106e34b7d64c> in <module>
2 print(np.diag(m.todense()).shape)
3
----> 4 m.diagonal()
~/.pyenv/versions/3.6.9/lib/python3.6/site-packages/scipy/sparse/compressed.py in diagonal(self, k)
531 rows, cols = self.shape
532 if k <= -rows or k >= cols:
--> 533 raise ValueError("k exceeds matrix dimensions")
534 fn = getattr(_sparsetools, self.format + "_diagonal")
535 y = np.empty(min(rows + min(k, 0), cols - max(k, 0)),
ValueError: k exceeds matrix dimensions
```
Output of the loop in the "complete" example:
```
bsr_matrix: ValueError('k exceeds matrix dimensions',)
coo_matrix: ValueError('k exceeds matrix dimensions',)
csc_matrix: ValueError('k exceeds matrix dimensions',)
csr_matrix: ValueError('k exceeds matrix dimensions',)
dia_matrix: ValueError('k exceeds matrix dimensions',)
dok_matrix: ValueError('k exceeds matrix dimensions',)
lil_matrix: ValueError('k exceeds matrix dimensions',)
```
#### Scipy/Numpy/Python version information:
```
1.4.1 1.17.4 sys.version_info(major=3, minor=6, micro=9, releaselevel='final', serial=0)
``` | 1.0 | ValueError 'k exceeds matrix dimensions' for sparse.diagonal() when 0 in sparse.shape - When a sparse matrix has a 0 in its shape, such as `(0, 0)`, `(0, 1)` or `(1, 0)`, calling `diagonal()` fails. This differs to `np.diag` on the equivalent dense array, which succeeds, returning an empty array.
The best behaviour here seems like it'd be open for debate, but it's unfortunate that the default `diagonal()` method doesn't work on every square sparse matrix. It can require adding special cases/conditionals around `.diagonal()` calls, such as https://github.com/stellargraph/stellargraph/pull/1378.
#### Reproducing code example:
Minimal:
```python
import scipy.sparse as sps
import numpy as np
m = sps.csr_matrix((0, 0))
print(np.diag(m.todense()).shape) # (0,)
m.diagonal() # ValueError: k exceeds matrix dimensions
```
"Complete" tests:
```python
import scipy.sparse as sps
# check all the sparse matrix classes
classes = [
sps.bsr_matrix,
sps.coo_matrix,
sps.csc_matrix,
sps.csr_matrix,
sps.dia_matrix,
sps.dok_matrix,
sps.lil_matrix,
]
for cls in classes:
try:
cls((0,0)).diagonal()
except Exception as e:
msg = e
else:
msg = None
print(f"{cls.__name__}: {msg!r}")
# For completeness, non-(0, 0) empty matrices
m = sps.csr_matrix((1, 0))
m.diagonal() # ValueError: k exceeds matrix dimensions
m = sps.csr_matrix((0, 1))
m.diagonal() # ValueError: k exceeds matrix dimensions
```
#### Error message:
Minimal example:
```
(0,)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-44-106e34b7d64c> in <module>
2 print(np.diag(m.todense()).shape)
3
----> 4 m.diagonal()
~/.pyenv/versions/3.6.9/lib/python3.6/site-packages/scipy/sparse/compressed.py in diagonal(self, k)
531 rows, cols = self.shape
532 if k <= -rows or k >= cols:
--> 533 raise ValueError("k exceeds matrix dimensions")
534 fn = getattr(_sparsetools, self.format + "_diagonal")
535 y = np.empty(min(rows + min(k, 0), cols - max(k, 0)),
ValueError: k exceeds matrix dimensions
```
Output of the loop in the "complete" example:
```
bsr_matrix: ValueError('k exceeds matrix dimensions',)
coo_matrix: ValueError('k exceeds matrix dimensions',)
csc_matrix: ValueError('k exceeds matrix dimensions',)
csr_matrix: ValueError('k exceeds matrix dimensions',)
dia_matrix: ValueError('k exceeds matrix dimensions',)
dok_matrix: ValueError('k exceeds matrix dimensions',)
lil_matrix: ValueError('k exceeds matrix dimensions',)
```
#### Scipy/Numpy/Python version information:
```
1.4.1 1.17.4 sys.version_info(major=3, minor=6, micro=9, releaselevel='final', serial=0)
``` | defect | valueerror k exceeds matrix dimensions for sparse diagonal when in sparse shape when a sparse matrix has a in its shape such as or calling diagonal fails this differs to np diag on the equivalent dense array which succeeds returning an empty array the best behaviour here seems like it d be open for debate but it s unfortunate that the default diagonal method doesn t work on every square sparse matrix it can require adding special cases conditionals around diagonal calls such as reproducing code example minimal python import scipy sparse as sps import numpy as np m sps csr matrix print np diag m todense shape m diagonal valueerror k exceeds matrix dimensions complete tests python import scipy sparse as sps check all the sparse matrix classes classes sps bsr matrix sps coo matrix sps csc matrix sps csr matrix sps dia matrix sps dok matrix sps lil matrix for cls in classes try cls diagonal except exception as e msg e else msg none print f cls name msg r for completeness non empty matrices m sps csr matrix m diagonal valueerror k exceeds matrix dimensions m sps csr matrix m diagonal valueerror k exceeds matrix dimensions error message minimal example valueerror traceback most recent call last in print np diag m todense shape m diagonal pyenv versions lib site packages scipy sparse compressed py in diagonal self k rows cols self shape if k cols raise valueerror k exceeds matrix dimensions fn getattr sparsetools self format diagonal y np empty min rows min k cols max k valueerror k exceeds matrix dimensions output of the loop in the complete example bsr matrix valueerror k exceeds matrix dimensions coo matrix valueerror k exceeds matrix dimensions csc matrix valueerror k exceeds matrix dimensions csr matrix valueerror k exceeds matrix dimensions dia matrix valueerror k exceeds matrix dimensions dok matrix valueerror k exceeds matrix dimensions lil matrix valueerror k exceeds matrix dimensions scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
60,215 | 17,023,371,236 | IssuesEvent | 2021-07-03 01:40:36 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Render railway in lower zoom level | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 8.35am, Friday, 27th February 2009]**
The main osm map does not render railways in zoom level < 8. Railway is an important way type and is the trunk of transportation network in many countries. I suppose main railway lines should be as important as trunk highways. Please consider render railways in lower zoom levels, at level z=5. | 1.0 | Render railway in lower zoom level - **[Submitted to the original trac issue database at 8.35am, Friday, 27th February 2009]**
The main osm map does not render railways in zoom level < 8. Railway is an important way type and is the trunk of transportation network in many countries. I suppose main railway lines should be as important as trunk highways. Please consider render railways in lower zoom levels, at level z=5. | defect | render railway in lower zoom level the main osm map does not render railways in zoom level railway is an important way type and is the trunk of transportation network in many countries i suppose main railway lines should be as important as trunk highways please consider render railways in lower zoom levels at level z | 1 |
390,671 | 26,868,626,962 | IssuesEvent | 2023-02-04 06:39:23 | module-federation/universe | https://api.github.com/repos/module-federation/universe | closed | Move "shared" into package utils | documentation enhancement good first issue help wanted no-issue-activity | Move the federated catchAll matcher code into nextjs-mf as part of the plugin utility chain
currently exists under demo/shared, should be moved to nextjs-mf/src/utils
Needs to be documented as well | 1.0 | Move "shared" into package utils - Move the federated catchAll matcher code into nextjs-mf as part of the plugin utility chain
currently exists under demo/shared, should be moved to nextjs-mf/src/utils
Needs to be documented as well | non_defect | move shared into package utils move the federated catchall matcher code into nextjs mf as part of the plugin utility chain currently exists under demo shared should be moved to nextjs mf src utils needs to be documented as well | 0 |
139,502 | 5,377,014,094 | IssuesEvent | 2017-02-23 10:48:13 | LearningLocker/learninglocker | https://api.github.com/repos/LearningLocker/learninglocker | closed | Failed to connect to: localhost:27017 | priority:low status:unconfirmed type:question | {{question}}?
This works
`tom@owly:~/src/webapps/php/learninglocker$ mongo --port 27017 -u "tom" -p "blafasel" --authenticationDatabase "lrs"
MongoDB shell version: 2.6.10
connecting to: 127.0.0.1:27017/test
> use lrs
switched to db lrs
> db.lrs.insert({"bla" : "fasel"})
WriteResult({ "nInserted" : 1 })
`
this is my app/config/local/database.php:
`<?php
return [
'connections' => [
'mongodb' => [
'driver' => 'mongodb',
'host' => 'localhost',
'port' => 27017,
'username' => 'tom',
'password' => 'REDACTED',
'database' => 'lrs' // Default name (removing this makes Travis fail).
],
],
];
`
when I try to use the command:
/usr/bin/php5.6 artisan migrate
I got:
Status: 500 Error: Failed to connect to: localhost:27017: Connection refusedSegmentation fault (core dumped)
Any hints for me, how to debug this? Installation process with composer worked without an error.
I'm on an Ubuntu 16.04 with default mongo 3.4 installation. PHP is 5.6 from PPA. Code is a current clone from here.
Regards
Tom
| 1.0 | Failed to connect to: localhost:27017 - {{question}}?
This works
`tom@owly:~/src/webapps/php/learninglocker$ mongo --port 27017 -u "tom" -p "blafasel" --authenticationDatabase "lrs"
MongoDB shell version: 2.6.10
connecting to: 127.0.0.1:27017/test
> use lrs
switched to db lrs
> db.lrs.insert({"bla" : "fasel"})
WriteResult({ "nInserted" : 1 })
`
this is my app/config/local/database.php:
`<?php
return [
'connections' => [
'mongodb' => [
'driver' => 'mongodb',
'host' => 'localhost',
'port' => 27017,
'username' => 'tom',
'password' => 'REDACTED',
'database' => 'lrs' // Default name (removing this makes Travis fail).
],
],
];
`
when I try to use the command:
/usr/bin/php5.6 artisan migrate
I got:
Status: 500 Error: Failed to connect to: localhost:27017: Connection refusedSegmentation fault (core dumped)
Any hints for me, how to debug this? Installation process with composer worked without an error.
I'm on an Ubuntu 16.04 with default mongo 3.4 installation. PHP is 5.6 from PPA. Code is a current clone from here.
Regards
Tom
| non_defect | failed to connect to localhost question this works tom owly src webapps php learninglocker mongo port u tom p blafasel authenticationdatabase lrs mongodb shell version connecting to test use lrs switched to db lrs db lrs insert bla fasel writeresult ninserted this is my app config local database php php return connections mongodb driver mongodb host localhost port username tom password redacted database lrs default name removing this makes travis fail when i try to use the command usr bin artisan migrate i got status error failed to connect to localhost connection refusedsegmentation fault core dumped any hints for me how to debug this installation process with composer worked without an error i m on an ubuntu with default mongo installation php is from ppa code is a current clone from here regards tom | 0 |
585,329 | 17,485,021,987 | IssuesEvent | 2021-08-09 09:52:02 | DostEducation/RP_IVR_analytics | https://api.github.com/repos/DostEducation/RP_IVR_analytics | closed | Create analytics dashboard for UNICEF Project | priority: high | - [ ] Create health metric for UNICEF project.
- [ ] Create Funnel metrics for UNICEF project. | 1.0 | Create analytics dashboard for UNICEF Project - - [ ] Create health metric for UNICEF project.
- [ ] Create Funnel metrics for UNICEF project. | non_defect | create analytics dashboard for unicef project create health metric for unicef project create funnel metrics for unicef project | 0 |
78,682 | 27,711,166,646 | IssuesEvent | 2023-03-14 14:20:25 | dotCMS/core | https://api.github.com/repos/dotCMS/core | reopened | Lucene queries for personas returning unexpected results | Type : Defect Needs Work Merged QA : Passed Internal Starter Generation Needed LTS: Priority Team : Falcon Release : 23.03 OKR : Customer Support | **Describe the bug**
Lucene queries for personas returning unexpected results
Reproduced on: 22.03.2, 22.08
Related ticket: https://dotcms.zendesk.com/agent/tickets/108233
**To Reproduce**
From a demo starter
1. Create a new content type called `Test CT` with a title field and a tag field called `testTags` that is indexed
2. Add a contentlet with the persona tag `ecoenthusiast`
3. Try a lucene query
Example 1
```
+contentType:TestCt
+TestCt.testTags("ecoenthusiast")
```
Example 2
```
+contentType:TestCt
+TestCt.testTags("ecoenthusiast:persona")
```
The first time I tried to recreate this only example 2 returned results. The second time, neither are returning results.
**Expected behavior**
The query should not require :persona appended, this was not required in previous versions. You should only need the name of the persona.
Beyond that, the results seem generally inconsistent. It might have something to do with how we are dealing with `:` in the query.
| 1.0 | Lucene queries for personas returning unexpected results - **Describe the bug**
Lucene queries for personas returning unexpected results
Reproduced on: 22.03.2, 22.08
Related ticket: https://dotcms.zendesk.com/agent/tickets/108233
**To Reproduce**
From a demo starter
1. Create a new content type called `Test CT` with a title field and a tag field called `testTags` that is indexed
2. Add a contentlet with the persona tag `ecoenthusiast`
3. Try a lucene query
Example 1
```
+contentType:TestCt
+TestCt.testTags("ecoenthusiast")
```
Example 2
```
+contentType:TestCt
+TestCt.testTags("ecoenthusiast:persona")
```
The first time I tried to recreate this only example 2 returned results. The second time, neither are returning results.
**Expected behavior**
The query should not require :persona appended, this was not required in previous versions. You should only need the name of the persona.
Beyond that, the results seem generally inconsistent. It might have something to do with how we are dealing with `:` in the query.
| defect | lucene queries for personas returning unexpected results describe the bug lucene queries for personas returning unexpected results reproduced on related ticket to reproduce from a demo starter create a new content type called test ct with a title field and a tag field called testtags that is indexed add a contentlet with the persona tag ecoenthusiast try a lucene query example contenttype testct testct testtags ecoenthusiast example contenttype testct testct testtags ecoenthusiast persona the first time i tried to recreate this only example returned results the second time neither are returning results expected behavior the query should not require persona appended this was not required in previous versions you should only need the name of the persona beyond that the results seem generally inconsistent it might have something to do with how we are dealing with in the query | 1 |
79,636 | 15,586,192,100 | IssuesEvent | 2021-03-18 01:22:56 | ziednov007/JavaSpring | https://api.github.com/repos/ziednov007/JavaSpring | opened | CVE-2020-24616 (High) detected in jackson-databind-2.9.6.jar | security vulnerability | ## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: JavaSpring/app/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.5.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-24616 (High) detected in jackson-databind-2.9.6.jar - ## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: JavaSpring/app/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.5.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file javaspring app build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
246,182 | 26,600,353,610 | IssuesEvent | 2023-01-23 15:21:33 | lukebrogan-mend/django.nV | https://api.github.com/repos/lukebrogan-mend/django.nV | closed | CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.js - autoclosed | security vulnerability | ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /taskManager/static/taskManager/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.2.0.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/django.nV/commit/442c6c7076c373c9762f875ec09227c88ad5d198">442c6c7076c373c9762f875ec09227c88ad5d198</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
| True | CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.js - autoclosed - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /taskManager/static/taskManager/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.2.0.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/django.nV/commit/442c6c7076c373c9762f875ec09227c88ad5d198">442c6c7076c373c9762f875ec09227c88ad5d198</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
| non_defect | cve medium detected in bootstrap js autoclosed cve medium severity vulnerability vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library taskmanager static taskmanager js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass | 0 |
121,210 | 25,940,083,084 | IssuesEvent | 2022-12-16 17:29:02 | microsoft/vscode-remote-release | https://api.github.com/repos/microsoft/vscode-remote-release | closed | VS Code Server Linux service | code-server | It would be nice to have a `systemd` service for the server, as this would allow it to be restarted automatically after a reboot, for instance. | 1.0 | VS Code Server Linux service - It would be nice to have a `systemd` service for the server, as this would allow it to be restarted automatically after a reboot, for instance. | non_defect | vs code server linux service it would be nice to have a systemd service for the server as this would allow it to be restarted automatically after a reboot for instance | 0 |
24,503 | 6,547,803,032 | IssuesEvent | 2017-09-04 16:35:01 | CentralProgramming/CentralProgramming.github.io | https://api.github.com/repos/CentralProgramming/CentralProgramming.github.io | closed | Team section cards aren't same size | code issues | The "you" card in the team section isn't the same size as the others. And actually, you can't notice it but the others are only the same size because there is the same amount of content. I think I know how to fix this though. | 1.0 | Team section cards aren't same size - The "you" card in the team section isn't the same size as the others. And actually, you can't notice it but the others are only the same size because there is the same amount of content. I think I know how to fix this though. | non_defect | team section cards aren t same size the you card in the team section isn t the same size as the others and actually you can t notice it but the others are only the same size because there is the same amount of content i think i know how to fix this though | 0 |
10,960 | 2,622,855,833 | IssuesEvent | 2015-03-04 08:07:42 | max99x/dict-lookup-chrome-ext | https://api.github.com/repos/max99x/dict-lookup-chrome-ext | closed | isInsideFrame() fails on scrolled pages with fixed frame. | auto-migrated Priority-Medium Type-Defect | ```
When doing a manual query (through the shortcut), the resulting form is
displayed in a fixed position in the center of the screen. If the containing
page is scrolled, isInsideFrame() does not calculate the frame position
correctly.
```
Original issue reported on code.google.com by `max99x` on 11 Mar 2010 at 11:35 | 1.0 | isInsideFrame() fails on scrolled pages with fixed frame. - ```
When doing a manual query (through the shortcut), the resulting form is
displayed in a fixed position in the center of the screen. If the containing
page is scrolled, isInsideFrame() does not calculate the frame position
correctly.
```
Original issue reported on code.google.com by `max99x` on 11 Mar 2010 at 11:35 | defect | isinsideframe fails on scrolled pages with fixed frame when doing a manual query through the shortcut the resulting form is displayed in a fixed position in the center of the screen if the containing page is scrolled isinsideframe does not calculate the frame position correctly original issue reported on code google com by on mar at | 1 |
32,896 | 6,965,081,079 | IssuesEvent | 2017-12-09 01:35:12 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Programmatic Menu - submenus don't expand and collapse when toggleable="true" | defect | This is a continuation of the issue found in #2344. While the ClassCastException no longer occurs, the actual functionality of toggling a programmatic menu's submenu does not work.
## 1) Environment
- PrimeFaces version: latest - 6.1.x, after applying the fix in #2347
- Does it work on the newest released PrimeFaces version? No. Version? Works in 5.3.0
- Does it work on the newest sources in GitHub? No (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source)
- Application server + version: All
- Affected browsers: All
## 2) Expected behavior
When toggling a submenu for a programmatic menu, the menu items should collapse and expand.
...
## 3) Actual behavior
MenuItems under a submenu do not toggle when using a menuModel
..
## 4) Steps to reproduce
Create a Programmatic menu using p:menu, set toggleable = true, and set model to a MenuModel defined in the controller
Also make sure you have the fix for issue #2344, #2347 loaded.
Then try toggling your menu.
Previously you would get a ClassCastException just loading the page, which commit #2347 fixed, however now the toggle doesn't work.
..
## 5) Sample XHTML
You can use PrimeFaces Showcase:
https://www.primefaces.org/showcase/ui/menu/menu.xhtml
But change line 36 from:
`<p:menu model="#{menuView.model}" />`
to
`<p:menu toggleable="true" model="#{menuView.model}" />`
..
## 6) Sample bean
Use PrimeFaces showcase MenuView.java
https://www.primefaces.org/showcase/ui/menu/menu.xhtml
..
## 7) More Investigation
I believe this is because in #1485, the menu.js was also modified to only hide or show elements with the .ui-submenu-child class.
Menu.js Line 779: `items.filter('.ui-submenu-child').hide();`
Menu.js Line 793: `items.filter('.ui-submenu-child').show();`
Because the .ui-submenu-child is not being added to the menuItem when using a menu model, none of the menuItems are being found to show or hide.
I'm not familiar enough with PrimeFaces code to suggest a resolution that doesn't break the issue #1485 fixed, so I would welcome suggestions. | 1.0 | Programmatic Menu - submenus don't expand and collapse when toggleable="true" - This is a continuation of the issue found in #2344. While the ClassCastException no longer occurs, the actual functionality of toggling a programmatic menu's submenu does not work.
## 1) Environment
- PrimeFaces version: latest - 6.1.x, after applying the fix in #2347
- Does it work on the newest released PrimeFaces version? No. Version? Works in 5.3.0
- Does it work on the newest sources in GitHub? No (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source)
- Application server + version: All
- Affected browsers: All
## 2) Expected behavior
When toggling a submenu for a programmatic menu, the menu items should collapse and expand.
...
## 3) Actual behavior
MenuItems under a submenu do not toggle when using a menuModel
..
## 4) Steps to reproduce
Create a Programmatic menu using p:menu, set toggleable = true, and set model to a MenuModel defined in the controller
Also make sure you have the fix for issue #2344, #2347 loaded.
Then try toggling your menu.
Previously you would get a ClassCastException just loading the page, which commit #2347 fixed, however now the toggle doesn't work.
..
## 5) Sample XHTML
You can use PrimeFaces Showcase:
https://www.primefaces.org/showcase/ui/menu/menu.xhtml
But change line 36 from:
`<p:menu model="#{menuView.model}" />`
to
`<p:menu toggleable="true" model="#{menuView.model}" />`
..
## 6) Sample bean
Use PrimeFaces showcase MenuView.java
https://www.primefaces.org/showcase/ui/menu/menu.xhtml
..
## 7) More Investigation
I believe this is because in #1485, the menu.js was also modified to only hide or show elements with the .ui-submenu-child class.
Menu.js Line 779: `items.filter('.ui-submenu-child').hide();`
Menu.js Line 793: `items.filter('.ui-submenu-child').show();`
Because the .ui-submenu-child is not being added to the menuItem when using a menu model, none of the menuItems are being found to show or hide.
I'm not familiar enough with PrimeFaces code to suggest a resolution that doesn't break the issue #1485 fixed, so I would welcome suggestions. | defect | programmatic menu submenus don t expand and collapse when toggleable true this is a continuation of the issue found in while the classcastexception no longer occurs the actual functionality of toggling a programmatic menu s submenu does not work environment primefaces version latest x after applying the fix in does it work on the newest released primefaces version no version works in does it work on the newest sources in github no build by source application server version all affected browsers all expected behavior when toggling a submenu for a programmatic menu the menu items should collapse and expand actual behavior menuitems under a submenu do not toggle when using a menumodel steps to reproduce create a programmatic menu using p menu set toggleable true and set model to a menumodel defined in the controller also make sure you have the fix for issue loaded then try toggling your menu previously you would get a classcastexception just loading the page which commit fixed however now the toggle doesn t work sample xhtml you can use primefaces showcase but change line from to sample bean use primefaces showcase menuview java more investigation i believe this is because in the menu js was also modified to only hide or show elements with the ui submenu child class menu js line items filter ui submenu child hide menu js line items filter ui submenu child show because the ui submenu child is not being added to the menuitem when using a menu model none of the menuitems are being found to show or hide i m not familiar enough with primefaces code to suggest a resolution that doesn t break the issue fixed so i would welcome suggestions | 1 |
897 | 2,594,286,561 | IssuesEvent | 2015-02-20 01:29:47 | BALL-Project/ball | https://api.github.com/repos/BALL-Project/ball | closed | Obsolete CONECT record format | C: BALL Core P: major R: fixed T: defect | **Reported by akdehof on 9 May 39563530 00:26 UTC**
Storing a pdb-file currently raises CONECT records for HET atoms in the format
CONECT 2 1 6 7 8 0 0 0 0 0 0
which seems to be the format as defined in 1992/1996:
COLUMNS DATA TYPE FIELD DEFINITION
1 - 6 Record name "CONECT"
7 - 11 Integer serial Atom serial number
12 - 16 Integer serial Serial number of bonded atom
17 - 21 Integer serial Serial number of bonded atom
22 - 26 Integer serial Serial number of bonded atom
27 - 31 Integer serial Serial number of bonded atom
32 - 36 Integer serial Serial number of hydrogen bonded
atom
37 - 41 Integer serial Serial number of hydrogen bonded
atom
42 - 46 Integer serial Serial number of salt bridged
atom
47 - 51 Integer serial Serial number of hydrogen bonded
atom
According to the current pdb-file documentation only four Integers for bonded atom-indices are allowed:
COLUMNS DATA TYPE FIELD DEFINITION
1 - 6 Record name "CONECT"
7 - 11 Integer serial Atom serial number
12 - 16 Integer serial Serial number of bonded atom
17 - 21 Integer serial Serial number of bonded atom
22 - 26 Integer serial Serial number of bonded atom
27 - 31 Integer serial Serial number of bonded atom
BALL/BALLView itself seems to be able to __read__ this old format but probably should __not__ __write__ into it. This causes problems when creating molecules that are later used by e.g. openbabel.
| 1.0 | Obsolete CONECT record format - **Reported by akdehof on 9 May 39563530 00:26 UTC**
Storing a pdb-file currently raises CONECT records for HET atoms in the format
CONECT 2 1 6 7 8 0 0 0 0 0 0
which seems to be the format as defined in 1992/1996:
COLUMNS DATA TYPE FIELD DEFINITION
1 - 6 Record name "CONECT"
7 - 11 Integer serial Atom serial number
12 - 16 Integer serial Serial number of bonded atom
17 - 21 Integer serial Serial number of bonded atom
22 - 26 Integer serial Serial number of bonded atom
27 - 31 Integer serial Serial number of bonded atom
32 - 36 Integer serial Serial number of hydrogen bonded
atom
37 - 41 Integer serial Serial number of hydrogen bonded
atom
42 - 46 Integer serial Serial number of salt bridged
atom
47 - 51 Integer serial Serial number of hydrogen bonded
atom
According to the current pdb-file documentation only four Integers for bonded atom-indices are allowed:
COLUMNS DATA TYPE FIELD DEFINITION
1 - 6 Record name "CONECT"
7 - 11 Integer serial Atom serial number
12 - 16 Integer serial Serial number of bonded atom
17 - 21 Integer serial Serial number of bonded atom
22 - 26 Integer serial Serial number of bonded atom
27 - 31 Integer serial Serial number of bonded atom
BALL/BALLView itself seems to be able to __read__ this old format but probably should __not__ __write__ into it. This causes problems when creating molecules that are later used by e.g. openbabel.
| defect | obsolete conect record format reported by akdehof on may utc storing a pdb file currently raises conect records for het atoms in the format conect which seems to be the format as defined in columns data type field definition record name conect integer serial atom serial number integer serial serial number of bonded atom integer serial serial number of bonded atom integer serial serial number of bonded atom integer serial serial number of bonded atom integer serial serial number of hydrogen bonded atom integer serial serial number of hydrogen bonded atom integer serial serial number of salt bridged atom integer serial serial number of hydrogen bonded atom according to the current pdb file documentation only four integers for bonded atom indices are allowed columns data type field definition record name conect integer serial atom serial number integer serial serial number of bonded atom integer serial serial number of bonded atom integer serial serial number of bonded atom integer serial serial number of bonded atom ball ballview itself seems to be able to read this old format but probably should not write into it this causes problems when creating molecules that are later used by e g openbabel | 1 |
65,759 | 12,683,765,834 | IssuesEvent | 2020-06-19 20:35:33 | postmanlabs/postman-app-support | https://api.github.com/repos/postmanlabs/postman-app-support | closed | Generated command for wget/curl incorrectly duplicates cookies | bug product/codegen | <!--
Please read through the [guidelines](https://github.com/postmanlabs/postman-app-support#guidelines-for-reporting-issues) before creating a new issue.
-->
**Describe the bug**
Exporting a request to replicate in `curl` or `wget` incorrectly duplicates cookies.
**To Reproduce**
Steps to reproduce the behavior:
1. Put in any URL for a GET request (ex: `https://google.com`)
2. Add a Header with the key `Cookie` and the value `blah`
3. Click Send more than once.
4. Attempt to get cURL code via `Code --> cURL`.
5. Observe that `blah` is included twice.
**Expected behavior**
The output of `Code --> cURL` should give a command that exactly replicates the Postman request. The cookie `blah` should only be included once.
**Actual Output (observe incorrect blah,blah)**
```
curl -X GET \
https://google.com \
-H 'Accept: */*' \
-H 'Accept-Encoding: gzip, deflate' \
-H 'Cache-Control: no-cache' \
-H 'Connection: keep-alive' \
-H 'Cookie: blah,blah; 1P_JAR=2020-03-11-22; NID=200=XRYon05k9-_FKVxUo52aHKEixBuXICYrdQ8wgw_rAQt4Me-s6VJMLh7JHPEPxjnfYaKw82iUwzF43lvVYpdGJUpmZgiG7LTfkq_7X4InAaEEeLdAFOB6H9xAlqbRBHEaLB7lMjl0BgQ8UKhjCHcYjGJWZr0FQZjcx3GEGZd2dmA' \
-H 'Postman-Token: 5b89950f-5e94-4d65-b733-2482233510b8,6aa23da2-b4d6-467f-b576-53ce135daf3d' \
-H 'Referer: https://google.com/' \
-H 'User-Agent: PostmanRuntime/7.18.0' \
-H 'cache-control: no-cache'
```
**App information (please complete the following information):**
- App Type: Native
- Postman Version: 7.9.0
- OS: macOS Catalina 10.15.3
**Additional context**
It took me a bunch of time to figure out why the cURL command was giving different results than Postman. This would be helpful to fix. :D
| 1.0 | Generated command for wget/curl incorrectly duplicates cookies - <!--
Please read through the [guidelines](https://github.com/postmanlabs/postman-app-support#guidelines-for-reporting-issues) before creating a new issue.
-->
**Describe the bug**
Exporting a request to replicate in `curl` or `wget` incorrectly duplicates cookies.
**To Reproduce**
Steps to reproduce the behavior:
1. Put in any URL for a GET request (ex: `https://google.com`)
2. Add a Header with the key `Cookie` and the value `blah`
3. Click Send more than once.
4. Attempt to get cURL code via `Code --> cURL`.
5. Observe that `blah` is included twice.
**Expected behavior**
The output of `Code --> cURL` should give a command that exactly replicates the Postman request. The cookie `blah` should only be included once.
**Actual Output (observe incorrect blah,blah)**
```
curl -X GET \
https://google.com \
-H 'Accept: */*' \
-H 'Accept-Encoding: gzip, deflate' \
-H 'Cache-Control: no-cache' \
-H 'Connection: keep-alive' \
-H 'Cookie: blah,blah; 1P_JAR=2020-03-11-22; NID=200=XRYon05k9-_FKVxUo52aHKEixBuXICYrdQ8wgw_rAQt4Me-s6VJMLh7JHPEPxjnfYaKw82iUwzF43lvVYpdGJUpmZgiG7LTfkq_7X4InAaEEeLdAFOB6H9xAlqbRBHEaLB7lMjl0BgQ8UKhjCHcYjGJWZr0FQZjcx3GEGZd2dmA' \
-H 'Postman-Token: 5b89950f-5e94-4d65-b733-2482233510b8,6aa23da2-b4d6-467f-b576-53ce135daf3d' \
-H 'Referer: https://google.com/' \
-H 'User-Agent: PostmanRuntime/7.18.0' \
-H 'cache-control: no-cache'
```
**App information (please complete the following information):**
- App Type: Native
- Postman Version: 7.9.0
- OS: macOS Catalina 10.15.3
**Additional context**
It took me a bunch of time to figure out why the cURL command was giving different results than Postman. This would be helpful to fix. :D
| non_defect | generated command for wget curl incorrectly duplicates cookies please read through the before creating a new issue describe the bug exporting a request to replicate in curl or wget incorrectly duplicates cookies to reproduce steps to reproduce the behavior put in any url for a get request ex add a header with the key cookie and the value blah click send more than once attempt to get curl code via code curl observe that blah is included twice expected behavior the output of code curl should give a command that exactly replicates the postman request the cookie blah should only be included once actual output observe incorrect blah blah curl x get h accept h accept encoding gzip deflate h cache control no cache h connection keep alive h cookie blah blah jar nid h postman token h referer h user agent postmanruntime h cache control no cache app information please complete the following information app type native postman version os macos catalina additional context it took me a bunch of time to figure out why the curl command was giving different results than postman this would be helpful to fix d | 0 |
9,947 | 2,616,013,674 | IssuesEvent | 2015-03-02 00:56:41 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | getMinerals returns 0 | auto-migrated Type-Defect | ```
What steps will reproduce the problem?
1. use bwapi 3.1.1 beta
2. run exampleAIclient.exe (the one that came with bwapi)
3. launch singleplayer custom game "astral balance"
What is the expected output? What do you see instead?
i expect to see the workers harvesting minerals, but they just sit there.
after adding code to display the size of the returned mineral and geyser sets,
they show 0. Tried with complete map info, still no result. tried calling
bwapiclient::update about 5 times, then testing mineral count, still zero.
What version of the product are you using? On what operating system?
3.1.1 beta on W7P and XP Pro sp3.
Please provide any additional information below.
```
Original issue reported on code.google.com by `dpershouse@gmail.com` on 21 Oct 2010 at 2:52 | 1.0 | getMinerals returns 0 - ```
What steps will reproduce the problem?
1. use bwapi 3.1.1 beta
2. run exampleAIclient.exe (the one that came with bwapi)
3. launch singleplayer custom game "astral balance"
What is the expected output? What do you see instead?
i expect to see the workers harvesting minerals, but they just sit there.
after adding code to display the size of the returned mineral and geyser sets,
they show 0. Tried with complete map info, still no result. tried calling
bwapiclient::update about 5 times, then testing mineral count, still zero.
What version of the product are you using? On what operating system?
3.1.1 beta on W7P and XP Pro sp3.
Please provide any additional information below.
```
Original issue reported on code.google.com by `dpershouse@gmail.com` on 21 Oct 2010 at 2:52 | defect | getminerals returns what steps will reproduce the problem use bwapi beta run exampleaiclient exe the one that came with bwapi launch singleplayer custom game astral balance what is the expected output what do you see instead i expect to see the workers harvesting minerals but they just sit there after adding code to display the size of the returned mineral and geyser sets they show tried with complete map info still no result tried calling bwapiclient update about times then testing mineral count still zero what version of the product are you using on what operating system beta on and xp pro please provide any additional information below original issue reported on code google com by dpershouse gmail com on oct at | 1 |
43,164 | 11,517,581,925 | IssuesEvent | 2020-02-14 08:43:20 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | PDF exporter downloads or pops open the file based on # of records in datatable | defect | Reported By PRO User;
> Observed with Primfaces 8 RC2 , PDF DataExporter.
> Issue:
> When datatable has more than 300 rows, data exporter is opening the pdf file on click of PDF exporter.
> When less number of rows(approx. 20) are present in data table, data exporter is downloading the pdf.
>Expected result:
>Always prefer the pdf to be downloaded, like in Primefaces 7, rather than opening, on click of PDF exporter icon. | 1.0 | PDF exporter downloads or pops open the file based on # of records in datatable - Reported By PRO User;
> Observed with Primfaces 8 RC2 , PDF DataExporter.
> Issue:
> When datatable has more than 300 rows, data exporter is opening the pdf file on click of PDF exporter.
> When less number of rows(approx. 20) are present in data table, data exporter is downloading the pdf.
>Expected result:
>Always prefer the pdf to be downloaded, like in Primefaces 7, rather than opening, on click of PDF exporter icon. | defect | pdf exporter downloads or pops open the file based on of records in datatable reported by pro user observed with primfaces pdf dataexporter issue when datatable has more than rows data exporter is opening the pdf file on click of pdf exporter when less number of rows approx are present in data table data exporter is downloading the pdf expected result always prefer the pdf to be downloaded like in primefaces rather than opening on click of pdf exporter icon | 1 |
65,066 | 19,087,802,577 | IssuesEvent | 2021-11-29 08:43:20 | cf-convention/cf-convention.github.io | https://api.github.com/repos/cf-convention/cf-convention.github.io | closed | Change Fran's affiliation in the information management team | defect | As announced in https://github.com/cf-convention/discuss/issues/138, implemented by https://github.com/cf-convention/cf-convention.github.io/pull/202, labelled as a defect so that it can be implemented quickly.
| 1.0 | Change Fran's affiliation in the information management team - As announced in https://github.com/cf-convention/discuss/issues/138, implemented by https://github.com/cf-convention/cf-convention.github.io/pull/202, labelled as a defect so that it can be implemented quickly.
| defect | change fran s affiliation in the information management team as announced in implemented by labelled as a defect so that it can be implemented quickly | 1 |
755,331 | 26,425,257,890 | IssuesEvent | 2023-01-14 04:19:14 | SuddenDevelopment/StopMotion | https://api.github.com/repos/SuddenDevelopment/StopMotion | closed | Add & Remove Space | Priority High | Should always be set to Add or Remove 1 space. Leaving **Set space** to set spacing values as needed.
| 1.0 | Add & Remove Space - Should always be set to Add or Remove 1 space. Leaving **Set space** to set spacing values as needed.
| non_defect | add remove space should always be set to add or remove space leaving set space to set spacing values as needed | 0 |
155,185 | 13,614,511,498 | IssuesEvent | 2020-09-23 13:20:58 | camunda/camunda-modeler | https://api.github.com/repos/camunda/camunda-modeler | opened | Update element-template documentation | documentation enhancement | __Is your feature request related to a problem? Please describe.__
With https://github.com/camunda/camunda-modeler/milestone/54 a couple of enhancements are introduced to element-templates. This enhancements shall be reflected in the [docs](https://github.com/camunda/camunda-modeler/tree/master/docs/element-templates)
__Describe the solution you'd like__
Update documentation to include:
- [ ] Updated screenshots (esp. element template search modal)
- [ ] Updated description regarding Apply/Unlink/Remove logic https://github.com/camunda/camunda-modeler/issues/1889
- [ ] Updated description regarding treating of non-present element-templates https://github.com/camunda/camunda-modeler/issues/1894
- [ ] Reference to cloud:connect plugin https://github.com/camunda/cloud-connect-modeler-plugin/issues/17
- [ ] Update bindings configuration documentation / usage of "default" type for `camunda:inputParamter` and `camunda:outputParameter` about:blank
__Describe alternatives you've considered__
n/a
__Additional context__
Add any other context or screenshots about the feature request here.
| 1.0 | Update element-template documentation - __Is your feature request related to a problem? Please describe.__
With https://github.com/camunda/camunda-modeler/milestone/54 a couple of enhancements are introduced to element-templates. This enhancements shall be reflected in the [docs](https://github.com/camunda/camunda-modeler/tree/master/docs/element-templates)
__Describe the solution you'd like__
Update documentation to include:
- [ ] Updated screenshots (esp. element template search modal)
- [ ] Updated description regarding Apply/Unlink/Remove logic https://github.com/camunda/camunda-modeler/issues/1889
- [ ] Updated description regarding treating of non-present element-templates https://github.com/camunda/camunda-modeler/issues/1894
- [ ] Reference to cloud:connect plugin https://github.com/camunda/cloud-connect-modeler-plugin/issues/17
- [ ] Update bindings configuration documentation / usage of "default" type for `camunda:inputParamter` and `camunda:outputParameter` about:blank
__Describe alternatives you've considered__
n/a
__Additional context__
Add any other context or screenshots about the feature request here.
| non_defect | update element template documentation is your feature request related to a problem please describe with a couple of enhancements are introduced to element templates this enhancements shall be reflected in the describe the solution you d like update documentation to include updated screenshots esp element template search modal updated description regarding apply unlink remove logic updated description regarding treating of non present element templates reference to cloud connect plugin update bindings configuration documentation usage of default type for camunda inputparamter and camunda outputparameter about blank describe alternatives you ve considered n a additional context add any other context or screenshots about the feature request here | 0 |
449,513 | 31,848,627,916 | IssuesEvent | 2023-09-14 22:22:39 | terraform-ibm-modules/terraform-ibm-cbr | https://api.github.com/repos/terraform-ibm-modules/terraform-ibm-cbr | opened | docs: After you create or update a rule, it might take up to 10 minutes for the rule to take effect. | documentation enhancement | ### Description
Add this statement in the documentation. | 1.0 | docs: After you create or update a rule, it might take up to 10 minutes for the rule to take effect. - ### Description
Add this statement in the documentation. | non_defect | docs after you create or update a rule it might take up to minutes for the rule to take effect description add this statement in the documentation | 0 |
81,933 | 31,817,258,751 | IssuesEvent | 2023-09-13 21:34:33 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | "New" Elements are Missing Tooltips | Type : Defect Merged QA : Needs Internal Team : Scout | [](https://mrkr.io/s/64de7430a209e725ae32cc71/0)
<img width="401" alt="image" src="https://github.com/dotCMS/core/assets/114953617/e33ab143-04f7-44be-93e3-d055a5b05eb2">
### Parent Issue
https://github.com/dotCMS/core/issues/25811
### Problem Statement
We're missing tooltips on the "new" elements.
I think those got left out of the original requirements, sorry!
Edit - Edit Page Content
Preview - Preview Page
Lock - Lock Page
What's Changed - Compare Changes
### Steps to Reproduce
### Acceptance Criteria
### dotCMS Version
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
### Assumptions & Initiation Needs
### Quality Assurance Notes & Workarounds
### Sub-Tasks & Estimates
---
**Reported by:** Damen Gilland (damen.gilland@dotcms.com)
**Source URL:** [https://dotcms-qa-master.dotcms.site/dotAdmin/#/edit-page/content?url=%2Fblog%2Fpost%2Fthe-ultimate-family-ski-trip-packing-list%3Fhost_id%3D48190c8c-42c4-46af-8d1a-0cd5db894797&language_id=1&host_id=48190c8c-42c4-46af-8d1a-0cd5db894797](https://dotcms-qa-master.dotcms.site/dotAdmin/#/edit-page/content?url=%2Fblog%2Fpost%2Fthe-ultimate-family-ski-trip-packing-list%3Fhost_id%3D48190c8c-42c4-46af-8d1a-0cd5db894797&language_id=1&host_id=48190c8c-42c4-46af-8d1a-0cd5db894797)
**Issue details:** [Open in Marker.io](https://app.marker.io/i/64de7430a209e725ae32cc74_8b986490744be5fb?advanced=1)
<table><tr><td><strong>Device type</strong></td><td>desktop</td></tr><tr><td><strong>Browser</strong></td><td>Firefox 116.0</td></tr><tr><td><strong>Screen Size</strong></td><td>3008 x 1692</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.15</td></tr><tr><td><strong>Viewport Size</strong></td><td>3008 x 1304</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@​2x</td></tr></table> | 1.0 | "New" Elements are Missing Tooltips - [](https://mrkr.io/s/64de7430a209e725ae32cc71/0)
<img width="401" alt="image" src="https://github.com/dotCMS/core/assets/114953617/e33ab143-04f7-44be-93e3-d055a5b05eb2">
### Parent Issue
https://github.com/dotCMS/core/issues/25811
### Problem Statement
We're missing tooltips on the "new" elements.
I think those got left out of the original requirements, sorry!
Edit - Edit Page Content
Preview - Preview Page
Lock - Lock Page
What's Changed - Compare Changes
### Steps to Reproduce
### Acceptance Criteria
### dotCMS Version
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
### Assumptions & Initiation Needs
### Quality Assurance Notes & Workarounds
### Sub-Tasks & Estimates
---
**Reported by:** Damen Gilland (damen.gilland@dotcms.com)
**Source URL:** [https://dotcms-qa-master.dotcms.site/dotAdmin/#/edit-page/content?url=%2Fblog%2Fpost%2Fthe-ultimate-family-ski-trip-packing-list%3Fhost_id%3D48190c8c-42c4-46af-8d1a-0cd5db894797&language_id=1&host_id=48190c8c-42c4-46af-8d1a-0cd5db894797](https://dotcms-qa-master.dotcms.site/dotAdmin/#/edit-page/content?url=%2Fblog%2Fpost%2Fthe-ultimate-family-ski-trip-packing-list%3Fhost_id%3D48190c8c-42c4-46af-8d1a-0cd5db894797&language_id=1&host_id=48190c8c-42c4-46af-8d1a-0cd5db894797)
**Issue details:** [Open in Marker.io](https://app.marker.io/i/64de7430a209e725ae32cc74_8b986490744be5fb?advanced=1)
<table><tr><td><strong>Device type</strong></td><td>desktop</td></tr><tr><td><strong>Browser</strong></td><td>Firefox 116.0</td></tr><tr><td><strong>Screen Size</strong></td><td>3008 x 1692</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.15</td></tr><tr><td><strong>Viewport Size</strong></td><td>3008 x 1304</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@​2x</td></tr></table> | defect | new elements are missing tooltips img width alt image src parent issue problem statement we re missing tooltips on the new elements i think those got left out of the original requirements sorry edit edit page content preview preview page lock lock page what s changed compare changes steps to reproduce acceptance criteria dotcms version external links slack conversations support tickets figma designs etc assumptions initiation needs quality assurance notes workarounds sub tasks estimates reported by damen gilland damen gilland dotcms com source url issue details device type desktop browser firefox screen size x os os x viewport size x zoom level pixel ratio | 1 |
78,841 | 27,780,915,170 | IssuesEvent | 2023-03-16 20:59:39 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | Data loss due to resilver after split brain | Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 11
Kernel Version | 5.10.0-21
Architecture | amd64
OpenZFS Version | 2.0.3-9
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
In a pool consisting of only one mirror of two drives, after a split brain situation, a resilver can overwrite the "more recent" copy.
My pool consists of a mirror of two disks.
Yesterday, disk 1 had too many I/O errors and was offlined by ZFS.
Today, disk 2 failed (disappeared from the ATA bus), suspending the pool.
I rebooted the server.
The pool came back up in degraded state with only disk 1 online (since disk 2 wasn't recognized by the system anymore).
I replugged disk 2. It was recognized again.
I onlined disk 2 again. Resilvering started to copy data from disk 1 to disk 2 without further warning. All data that had been written in that day to disk 2 was gone.
### Describe how to reproduce the problem
```
fallocate -l 256M /tmp/disk1
fallocate -l 256M /tmp/disk2
zpool create test mirror /tmp/disk[12]
date > /test/file1
zpool offline test /tmp/disk1
date > /test/file2
zpool export test
mkdir /tmp/offline
mv /tmp/disk2 /tmp/offline/
zpool import -d /tmp/disk1 test
mv /tmp/offline/disk2 /tmp/
zpool online test /tmp/disk2
ls /test
```
Although disk2 would have had file2 on it, it's gone now.
### Remarks
I don't have much knowledge about ZFS internals. Just looking at the output of `zdb -l`, this probably could have been detected by `zpool online` in two different ways:
1. The TXG of disk2 is greater than that of disk1. This could probably change, however, if I do enough stuff to disk1 while disk2 is still gone.
2. The vdev_tree on disk2 lists disk1 as offline, but a device with the same guid is online in the pool when onlining disk2.
I'm not even expecting some magic to happen to somehow merge the two disks. I just expect a (possible) split-brain scenario to be detected and a big fat warning to appear before resilvering over data. | 1.0 | Data loss due to resilver after split brain - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 11
Kernel Version | 5.10.0-21
Architecture | amd64
OpenZFS Version | 2.0.3-9
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
In a pool consisting of only one mirror of two drives, after a split brain situation, a resilver can overwrite the "more recent" copy.
My pool consists of a mirror of two disks.
Yesterday, disk 1 had too many I/O errors and was offlined by ZFS.
Today, disk 2 failed (disappeared from the ATA bus), suspending the pool.
I rebooted the server.
The pool came back up in degraded state with only disk 1 online (since disk 2 wasn't recognized by the system anymore).
I replugged disk 2. It was recognized again.
I onlined disk 2 again. Resilvering started to copy data from disk 1 to disk 2 without further warning. All data that had been written in that day to disk 2 was gone.
### Describe how to reproduce the problem
```
fallocate -l 256M /tmp/disk1
fallocate -l 256M /tmp/disk2
zpool create test mirror /tmp/disk[12]
date > /test/file1
zpool offline test /tmp/disk1
date > /test/file2
zpool export test
mkdir /tmp/offline
mv /tmp/disk2 /tmp/offline/
zpool import -d /tmp/disk1 test
mv /tmp/offline/disk2 /tmp/
zpool online test /tmp/disk2
ls /test
```
Although disk2 would have had file2 on it, it's gone now.
### Remarks
I don't have much knowledge about ZFS internals. Just looking at the output of `zdb -l`, this probably could have been detected by `zpool online` in two different ways:
1. The TXG of disk2 is greater than that of disk1. This could probably change, however, if I do enough stuff to disk1 while disk2 is still gone.
2. The vdev_tree on disk2 lists disk1 as offline, but a device with the same guid is online in the pool when onlining disk2.
I'm not even expecting some magic to happen to somehow merge the two disks. I just expect a (possible) split-brain scenario to be detected and a big fat warning to appear before resilvering over data. | defect | data loss due to resilver after split brain thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name debian distribution version kernel version architecture openzfs version command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing in a pool consisting of only one mirror of two drives after a split brain situation a resilver can overwrite the more recent copy my pool consists of a mirror of two disks yesterday disk had too many i o errors and was offlined by zfs today disk failed disappeared from the ata bus suspending the pool i rebooted the server the pool came back up in degraded state with only disk online since disk wasn t recognized by the system anymore i replugged disk it was recognized again i onlined disk again resilvering started to copy data from disk to disk without further warning all data that had been written in that day to disk was gone describe how to reproduce the problem fallocate l tmp fallocate l tmp zpool create test mirror tmp disk date test zpool offline test tmp date test zpool export test mkdir tmp offline mv tmp tmp offline zpool import d tmp test mv tmp offline tmp zpool online test tmp ls test although would have had on it it s gone now remarks i don t have much knowledge about zfs internals just looking at the output of zdb l this probably could have been detected by zpool online in two different ways the txg of is greater than that of this could probably change however if i do enough stuff to while is still gone the vdev tree on lists as offline but a device with the same guid is online in the pool when onlining i m not even expecting some magic to happen to somehow merge the two disks i just expect a possible split brain scenario to be detected and a big fat warning to appear before resilvering over data | 1 |
13,138 | 2,732,910,582 | IssuesEvent | 2015-04-17 10:10:14 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | Missing error message for faulty openIDs | auto-migrated Priority-High Type-Defect | ```
As-is:
When I try to log in with a faulty openID (e.g. http:// instead of https://
for the HPI-openID), there is no error message. Only a white page is displayed.
To-be:
There should be an error page indicating that something went wrong in the
login process.
```
Original issue reported on code.google.com by `gero.dec...@googlemail.com` on 11 Jun 2008 at 12:06 | 1.0 | Missing error message for faulty openIDs - ```
As-is:
When I try to log in with a faulty openID (e.g. http:// instead of https://
for the HPI-openID), there is no error message. Only a white page is displayed.
To-be:
There should be an error page indicating that something went wrong in the
login process.
```
Original issue reported on code.google.com by `gero.dec...@googlemail.com` on 11 Jun 2008 at 12:06 | defect | missing error message for faulty openids as is when i try to log in with a faulty openid e g http instead of https for the hpi openid there is no error message only a white page is displayed to be there should be an error page indicating that something went wrong in the login process original issue reported on code google com by gero dec googlemail com on jun at | 1 |
51,907 | 13,211,336,061 | IssuesEvent | 2020-08-15 22:23:58 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | DomTools release notes (Trac #1245) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1245">https://code.icecube.wisc.edu/projects/icecube/ticket/1245</a>, reported by kjmeagherand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"_ts": "1550067117911749",
"description": "no trunk section\nr136231 should be included in release notes",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2015-08-20T08:50:58",
"component": "combo reconstruction",
"summary": "DomTools release notes",
"priority": "critical",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | DomTools release notes (Trac #1245) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1245">https://code.icecube.wisc.edu/projects/icecube/ticket/1245</a>, reported by kjmeagherand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"_ts": "1550067117911749",
"description": "no trunk section\nr136231 should be included in release notes",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2015-08-20T08:50:58",
"component": "combo reconstruction",
"summary": "DomTools release notes",
"priority": "critical",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | domtools release notes trac migrated from json status closed changetime ts description no trunk section should be included in release notes reporter kjmeagher cc resolution fixed time component combo reconstruction summary domtools release notes priority critical keywords milestone owner olivas type defect | 1 |
212,671 | 16,492,693,413 | IssuesEvent | 2021-05-25 06:51:42 | Clariteia/minos_microservice_common | https://api.github.com/repos/Clariteia/minos_microservice_common | closed | Document Configuration package | documentation | Create Sphinx documentation por the configuration module. This should contain:
- Description of the public methods for autodoc.
- General description of the configuration object. | 1.0 | Document Configuration package - Create Sphinx documentation por the configuration module. This should contain:
- Description of the public methods for autodoc.
- General description of the configuration object. | non_defect | document configuration package create sphinx documentation por the configuration module this should contain description of the public methods for autodoc general description of the configuration object | 0 |
237,921 | 18,172,536,039 | IssuesEvent | 2021-09-27 21:47:36 | llighter/yunha-ind-astro | https://api.github.com/repos/llighter/yunha-ind-astro | closed | 이슈 등록을 위한 템플릿 만들기 | documentation enhancement | 새로운 포스트를 작성하거나 블로그 기능을 추가하는 작업에 대한 요구사항을 기록하고 진행상황을 추적하기 위해서는 이슈에 등록을 해야한다.
자주 사용할 이슈 양식을 템플릿으로 만들어서 관리해야한다.
기본적인 이슈 양식은 아래와 같다.
* 새로운 컨텐츠 작성
* 버그 리포트
* 기능 요청 | 1.0 | 이슈 등록을 위한 템플릿 만들기 - 새로운 포스트를 작성하거나 블로그 기능을 추가하는 작업에 대한 요구사항을 기록하고 진행상황을 추적하기 위해서는 이슈에 등록을 해야한다.
자주 사용할 이슈 양식을 템플릿으로 만들어서 관리해야한다.
기본적인 이슈 양식은 아래와 같다.
* 새로운 컨텐츠 작성
* 버그 리포트
* 기능 요청 | non_defect | 이슈 등록을 위한 템플릿 만들기 새로운 포스트를 작성하거나 블로그 기능을 추가하는 작업에 대한 요구사항을 기록하고 진행상황을 추적하기 위해서는 이슈에 등록을 해야한다 자주 사용할 이슈 양식을 템플릿으로 만들어서 관리해야한다 기본적인 이슈 양식은 아래와 같다 새로운 컨텐츠 작성 버그 리포트 기능 요청 | 0 |
18,156 | 3,029,733,191 | IssuesEvent | 2015-08-04 14:10:25 | lemire/lbimproved | https://api.github.com/repos/lemire/lbimproved | closed | url of the requirement for the Spatial Index Library cannot be reached | auto-migrated Priority-Medium Type-Defect | ```
Could you please update for a new link for the library? Thank you.
```
Original issue reported on code.google.com by `achm6...@gmail.com` on 8 Feb 2015 at 9:04 | 1.0 | url of the requirement for the Spatial Index Library cannot be reached - ```
Could you please update for a new link for the library? Thank you.
```
Original issue reported on code.google.com by `achm6...@gmail.com` on 8 Feb 2015 at 9:04 | defect | url of the requirement for the spatial index library cannot be reached could you please update for a new link for the library thank you original issue reported on code google com by gmail com on feb at | 1 |
742,944 | 25,879,396,252 | IssuesEvent | 2022-12-14 10:12:44 | biodiversitydata-se/biocollect | https://api.github.com/repos/biodiversitydata-se/biocollect | closed | sorting IWC with rank does not work (sort by total instead) | bug 3-Medium priority | From Fredrik, on IWC survey
`Här kommer en bild som illustrerar punkt 1. Jag har här sorterat efter rank och som du kan se blir det istället ”Total” som sorteras.`

| 1.0 | sorting IWC with rank does not work (sort by total instead) - From Fredrik, on IWC survey
`Här kommer en bild som illustrerar punkt 1. Jag har här sorterat efter rank och som du kan se blir det istället ”Total” som sorteras.`

| non_defect | sorting iwc with rank does not work sort by total instead from fredrik on iwc survey här kommer en bild som illustrerar punkt jag har här sorterat efter rank och som du kan se blir det istället ”total” som sorteras | 0 |
56,147 | 14,948,513,409 | IssuesEvent | 2021-01-26 10:11:02 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | DataTable: localDateTime field ignoring LocalDateTimeConverter | defect | @melloware @Rapster It seems that the field notation ignores existing faces converters.
I am seeing this for LocalDateTime fields.
```
<p:column field="startDate" headerText="startDate" />
```
delivers field values of the default LocalDateTime pattern, e.g. 2021-01-22T09:02:43.147
while in case of
```
<p:column field="startDate" headerText="startDate">
<h:outputText value="#{entity.startDate}">
</p:column>
```
the field values are actually converted using the existing converter, e.g. 2021-01-22 09:02.
my converter is defined as follows:
```
@FacesConverter(forClass = LocalDateTime.class)
public class LocalDateTimeConverter implements Converter<Object> {
...
private static String getPattern(UIComponent component) {
String pattern = (String) component.getAttributes().get("pattern");
if (pattern == null) {
pattern = "yyyy-MM-dd HH:mm";
}
return pattern;
}
}
```
I am not sure whether this a bug report or feature request. In any case, many thanks in advance for having a view on this issue. | 1.0 | DataTable: localDateTime field ignoring LocalDateTimeConverter - @melloware @Rapster It seems that the field notation ignores existing faces converters.
I am seeing this for LocalDateTime fields.
```
<p:column field="startDate" headerText="startDate" />
```
delivers field values of the default LocalDateTime pattern, e.g. 2021-01-22T09:02:43.147
while in case of
```
<p:column field="startDate" headerText="startDate">
<h:outputText value="#{entity.startDate}">
</p:column>
```
the field values are actually converted using the existing converter, e.g. 2021-01-22 09:02.
my converter is defined as follows:
```
@FacesConverter(forClass = LocalDateTime.class)
public class LocalDateTimeConverter implements Converter<Object> {
...
private static String getPattern(UIComponent component) {
String pattern = (String) component.getAttributes().get("pattern");
if (pattern == null) {
pattern = "yyyy-MM-dd HH:mm";
}
return pattern;
}
}
```
I am not sure whether this a bug report or feature request. In any case, many thanks in advance for having a view on this issue. | defect | datatable localdatetime field ignoring localdatetimeconverter melloware rapster it seems that the field notation ignores existing faces converters i am seeing this for localdatetime fields delivers field values of the default localdatetime pattern e g while in case of the field values are actually converted using the existing converter e g my converter is defined as follows facesconverter forclass localdatetime class public class localdatetimeconverter implements converter private static string getpattern uicomponent component string pattern string component getattributes get pattern if pattern null pattern yyyy mm dd hh mm return pattern i am not sure whether this a bug report or feature request in any case many thanks in advance for having a view on this issue | 1 |
267,138 | 20,190,908,645 | IssuesEvent | 2022-02-11 05:16:50 | openssl/openssl | https://api.github.com/repos/openssl/openssl | opened | Conflicting license Information in OpenSSL 3.0.1 codebase few files | issue: documentation |
Hi All,
Please note headers of few files in OpenSSL version 3.0.1 codebase, mention that the files are under OpenSSL license but the link referred in the file header mentions that the OpenSSL component 3.0 release, and later releases derived from that are under Apache License v2.
Hence, this is conflicting information, should these files be treated under Apache v2 license or OpenSSL license. The LICENSE.text file in OpenSSL 3.0.1 mentions Apache v2 license and majority of the files header information also mention that the files are under Apache v2 license but few files have this conflicting information.
Please find below the header information of these files and files paths:
Header:
* Licensed under the OpenSSL license (the "License"). You may not use
* this file except in compliance with the License. You can obtain a copy
* in the file LICENSE in the source distribution or at
* https://www.openssl.org/source/license.html
File path of these files:
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/f_impl64.c
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/f_impl.h
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/arch_intrinsics.h
openssl-openssl-3.0.1/include/crypto/pem.h
openssl-openssl-3.0.1/test/recipes/03-test_internal_sm3.t
openssl-openssl-3.0.1/test/dane-cross.in
openssl-openssl-3.0.1/crypto/ec/asm/ecp_nistp521-ppc64.pl
openssl-openssl-3.0.1/apps/lib/engine_loader.c
openssl-openssl-3.0.1/apps/include/engine_loader.h
openssl-openssl-3.0.1/doc/man1/openssl-cmp.pod.in
openssl-openssl-3.0.1/test/pbetest.c
openssl-openssl-3.0.1/test/sm3_internal_test.c
Please clarify the license of these files, should these be treated as under Apache 2.0 license or OpenSSL license, and change the header information if required for more clarity
Thank you
| 1.0 | Conflicting license Information in OpenSSL 3.0.1 codebase few files -
Hi All,
Please note headers of few files in OpenSSL version 3.0.1 codebase, mention that the files are under OpenSSL license but the link referred in the file header mentions that the OpenSSL component 3.0 release, and later releases derived from that are under Apache License v2.
Hence, this is conflicting information, should these files be treated under Apache v2 license or OpenSSL license. The LICENSE.text file in OpenSSL 3.0.1 mentions Apache v2 license and majority of the files header information also mention that the files are under Apache v2 license but few files have this conflicting information.
Please find below the header information of these files and files paths:
Header:
* Licensed under the OpenSSL license (the "License"). You may not use
* this file except in compliance with the License. You can obtain a copy
* in the file LICENSE in the source distribution or at
* https://www.openssl.org/source/license.html
File path of these files:
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/f_impl64.c
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/f_impl.h
openssl-openssl-3.0.1/crypto/ec/curve448/arch_64/arch_intrinsics.h
openssl-openssl-3.0.1/include/crypto/pem.h
openssl-openssl-3.0.1/test/recipes/03-test_internal_sm3.t
openssl-openssl-3.0.1/test/dane-cross.in
openssl-openssl-3.0.1/crypto/ec/asm/ecp_nistp521-ppc64.pl
openssl-openssl-3.0.1/apps/lib/engine_loader.c
openssl-openssl-3.0.1/apps/include/engine_loader.h
openssl-openssl-3.0.1/doc/man1/openssl-cmp.pod.in
openssl-openssl-3.0.1/test/pbetest.c
openssl-openssl-3.0.1/test/sm3_internal_test.c
Please clarify the license of these files, should these be treated as under Apache 2.0 license or OpenSSL license, and change the header information if required for more clarity
Thank you
| non_defect | conflicting license information in openssl codebase few files hi all please note headers of few files in openssl version codebase mention that the files are under openssl license but the link referred in the file header mentions that the openssl component release and later releases derived from that are under apache license hence this is conflicting information should these files be treated under apache license or openssl license the license text file in openssl mentions apache license and majority of the files header information also mention that the files are under apache license but few files have this conflicting information please find below the header information of these files and files paths header licensed under the openssl license the license you may not use this file except in compliance with the license you can obtain a copy in the file license in the source distribution or at file path of these files openssl openssl crypto ec arch f c openssl openssl crypto ec arch f impl h openssl openssl crypto ec arch arch intrinsics h openssl openssl include crypto pem h openssl openssl test recipes test internal t openssl openssl test dane cross in openssl openssl crypto ec asm ecp pl openssl openssl apps lib engine loader c openssl openssl apps include engine loader h openssl openssl doc openssl cmp pod in openssl openssl test pbetest c openssl openssl test internal test c please clarify the license of these files should these be treated as under apache license or openssl license and change the header information if required for more clarity thank you | 0 |
42,075 | 10,780,786,731 | IssuesEvent | 2019-11-04 13:43:31 | mozilla-lockwise/lockwise-android | https://api.github.com/repos/mozilla-lockwise/lockwise-android | opened | Entries can be saved without a username | feature-CUD 🐄 type: defect | ## Steps to reproduce
1. Launch Lockwise.
2. Login with valid credentials.
3. Access any account.
4. Enter the edit view.
5. Delete the data from `username`
### Expected behavior
Mandatory field error message is displayed for `username` field when trying to save it without a username.
### Actual behavior
Entries can be saved without a username
### Device & build information
* Device: **Google Pixel 3a XL(Android 10)**
* Build version: **3.1.0 (Build 5361)**
### Notes
Attachments:

| 1.0 | Entries can be saved without a username - ## Steps to reproduce
1. Launch Lockwise.
2. Login with valid credentials.
3. Access any account.
4. Enter the edit view.
5. Delete the data from `username`
### Expected behavior
Mandatory field error message is displayed for `username` field when trying to save it without a username.
### Actual behavior
Entries can be saved without a username
### Device & build information
* Device: **Google Pixel 3a XL(Android 10)**
* Build version: **3.1.0 (Build 5361)**
### Notes
Attachments:

| defect | entries can be saved without a username steps to reproduce launch lockwise login with valid credentials access any account enter the edit view delete the data from username expected behavior mandatory field error message is displayed for username field when trying to save it without a username actual behavior entries can be saved without a username device build information device google pixel xl android build version build notes attachments | 1 |
62,767 | 17,191,535,839 | IssuesEvent | 2021-07-16 11:43:33 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | closed | HVAC-Diagram.exe broken in v9.5 | AuxiliaryTool Defect | Issue overview
--------------
The HVAC-Diagram.exe that comes with V9.5 installation (official release) does not work properly. In nearly all simulations runs with EP-Launch, the HVAC schematic diagram will not be properly generated.
This was tested to be the case in two PCs with Windows 10 installations. A further check with command line shows that the HVAC-Diagram.exe for v9.5 is somehow broken---when it runs (with properly eplusout file in place), it does not give any output.
A temporary workaround is to copy the HVAC-Diagram.exe from v9.4 installation to place the one in v9.5. The version from v9.4 works fine, and it will also help generate the HVAC schematic diagrams properly for v9.5 runs.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): At least on Windows 10, both Pro and Home Editions
- Version of EnergyPlus (if using an intermediate build, include SHA): v9.5
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | HVAC-Diagram.exe broken in v9.5 - Issue overview
--------------
The HVAC-Diagram.exe that comes with V9.5 installation (official release) does not work properly. In nearly all simulations runs with EP-Launch, the HVAC schematic diagram will not be properly generated.
This was tested to be the case in two PCs with Windows 10 installations. A further check with command line shows that the HVAC-Diagram.exe for v9.5 is somehow broken---when it runs (with properly eplusout file in place), it does not give any output.
A temporary workaround is to copy the HVAC-Diagram.exe from v9.4 installation to place the one in v9.5. The version from v9.4 works fine, and it will also help generate the HVAC schematic diagrams properly for v9.5 runs.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): At least on Windows 10, both Pro and Home Editions
- Version of EnergyPlus (if using an intermediate build, include SHA): v9.5
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| defect | hvac diagram exe broken in issue overview the hvac diagram exe that comes with installation official release does not work properly in nearly all simulations runs with ep launch the hvac schematic diagram will not be properly generated this was tested to be the case in two pcs with windows installations a further check with command line shows that the hvac diagram exe for is somehow broken when it runs with properly eplusout file in place it does not give any output a temporary workaround is to copy the hvac diagram exe from installation to place the one in the version from works fine and it will also help generate the hvac schematic diagrams properly for runs details some additional details for this issue if relevant platform operating system version at least on windows both pro and home editions version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 1 |
546,199 | 16,006,750,574 | IssuesEvent | 2021-04-20 04:32:30 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | {Accounts] Upgrade to Dev Tier doesn't update the additional amount paid in the Admin backend | Category: Accounts Category: Web Priority: High Squad: Lattice Mushroom Status: Fixed | Reproduce:
1. Log in or create an account to buy Eco
2. Account in db should update the amount paid to $30.00
3. Upgrade the same account to any tier.
4. Account in db still stays at $30.00 | 1.0 | {Accounts] Upgrade to Dev Tier doesn't update the additional amount paid in the Admin backend - Reproduce:
1. Log in or create an account to buy Eco
2. Account in db should update the amount paid to $30.00
3. Upgrade the same account to any tier.
4. Account in db still stays at $30.00 | non_defect | accounts upgrade to dev tier doesn t update the additional amount paid in the admin backend reproduce log in or create an account to buy eco account in db should update the amount paid to upgrade the same account to any tier account in db still stays at | 0 |
18,011 | 3,016,176,143 | IssuesEvent | 2015-07-30 00:00:14 | catmaid/CATMAID | https://api.github.com/repos/catmaid/CATMAID | opened | Treenode connectors should be unique | priority: important type: defect | `treenode_connector` should have a unique constraint on the tuple: `project_id`, `treenode_id`, `connector_id`, `relation_id`
In my old database there are 130 duplicates for this criterion. This could impact connectivity analysis and the ability to implement, e.g., #1042 robustly.
The migration to deduplicate these should use the oldest creation time (and creator) and the newest edition time. | 1.0 | Treenode connectors should be unique - `treenode_connector` should have a unique constraint on the tuple: `project_id`, `treenode_id`, `connector_id`, `relation_id`
In my old database there are 130 duplicates for this criterion. This could impact connectivity analysis and the ability to implement, e.g., #1042 robustly.
The migration to deduplicate these should use the oldest creation time (and creator) and the newest edition time. | defect | treenode connectors should be unique treenode connector should have a unique constraint on the tuple project id treenode id connector id relation id in my old database there are duplicates for this criterion this could impact connectivity analysis and the ability to implement e g robustly the migration to deduplicate these should use the oldest creation time and creator and the newest edition time | 1 |
48,686 | 13,184,718,283 | IssuesEvent | 2020-08-12 19:58:04 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | ROOT version detection. (Trac #58) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/58
, reported by blaufuss and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-08-23T23:19:48",
"description": "PORT tools offer two versions of ROOT. (5.10.00 and 5.14.00b currently)\nbut cmake is only setup to deal with 5.10.00. Should\neither:\n\n-use user-set ROOTSYS to pick on or the other\nor\n-try both available (probably try newer version first? Like\nmysql does)",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1187911188000000",
"component": "cmake",
"summary": "ROOT version detection.",
"priority": "normal",
"keywords": "",
"time": "2007-06-11T14:40:30",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | ROOT version detection. (Trac #58) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/58
, reported by blaufuss and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-08-23T23:19:48",
"description": "PORT tools offer two versions of ROOT. (5.10.00 and 5.14.00b currently)\nbut cmake is only setup to deal with 5.10.00. Should\neither:\n\n-use user-set ROOTSYS to pick on or the other\nor\n-try both available (probably try newer version first? Like\nmysql does)",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1187911188000000",
"component": "cmake",
"summary": "ROOT version detection.",
"priority": "normal",
"keywords": "",
"time": "2007-06-11T14:40:30",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | root version detection trac migrated from reported by blaufuss and owned by troy json status closed changetime description port tools offer two versions of root and currently nbut cmake is only setup to deal with should neither n n use user set rootsys to pick on or the other nor n try both available probably try newer version first like nmysql does reporter blaufuss cc resolution fixed ts component cmake summary root version detection priority normal keywords time milestone owner troy type defect | 1 |
65,991 | 19,848,433,632 | IssuesEvent | 2022-01-21 09:33:22 | scipy/scipy | https://api.github.com/repos/scipy/scipy | opened | BUG: Documentation of stft misses the import statement for numpy | defect | ### Describe your issue.
The sample code given for stft in the signal module misses the following line:
import numpy as np
### Reproducing Code Example
```python
from scipy import signal
import matplotlib.pyplot as plt
rng = np.random.default_rng()
```
### Error message
```shell
NameError: name 'np' is not defined
```
### SciPy/NumPy/Python version information
1.7.3 1.21.4 sys.version_info(major=3, minor=9, micro=9, releaselevel='final', serial=0) | 1.0 | BUG: Documentation of stft misses the import statement for numpy - ### Describe your issue.
The sample code given for stft in the signal module misses the following line:
import numpy as np
### Reproducing Code Example
```python
from scipy import signal
import matplotlib.pyplot as plt
rng = np.random.default_rng()
```
### Error message
```shell
NameError: name 'np' is not defined
```
### SciPy/NumPy/Python version information
1.7.3 1.21.4 sys.version_info(major=3, minor=9, micro=9, releaselevel='final', serial=0) | defect | bug documentation of stft misses the import statement for numpy describe your issue the sample code given for stft in the signal module misses the following line import numpy as np reproducing code example python from scipy import signal import matplotlib pyplot as plt rng np random default rng error message shell nameerror name np is not defined scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
138,711 | 11,212,194,200 | IssuesEvent | 2020-01-06 17:01:10 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | The 'Add new clause' shows strangely in Query Builder | :gear: tables :heavy_check_mark: duplicate 🌐 localization 🧪 testing | **Storage Explorer Version:** 1.11.2
**Build**: [20191220.9](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3338616)
**Branch**: hotfix/1.11.2-to-master
**Language**: German / Russian
**Platform/OS**: Windows 10/ CentOS 7.6.1810/ MacOS High Sierra
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch Storage Explorer.
2. Open 'Settings' -> Application (Regional Settings) -> Select 'Deutsche' -> Restart Storage Explorer.
3. Expand one storage account -> Tables -> Create a new table -> Open 'Query'.
4. Check the localized 'Add new clause'.
**Expect Experience:**
The ‘Add new clause’ shows in the same line.
**Actual Experience:**
1. The 'Add new clause' shows strangely.

2. ENU Screenshot.

| 1.0 | The 'Add new clause' shows strangely in Query Builder - **Storage Explorer Version:** 1.11.2
**Build**: [20191220.9](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3338616)
**Branch**: hotfix/1.11.2-to-master
**Language**: German / Russian
**Platform/OS**: Windows 10/ CentOS 7.6.1810/ MacOS High Sierra
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch Storage Explorer.
2. Open 'Settings' -> Application (Regional Settings) -> Select 'Deutsche' -> Restart Storage Explorer.
3. Expand one storage account -> Tables -> Create a new table -> Open 'Query'.
4. Check the localized 'Add new clause'.
**Expect Experience:**
The ‘Add new clause’ shows in the same line.
**Actual Experience:**
1. The 'Add new clause' shows strangely.

2. ENU Screenshot.

| non_defect | the add new clause shows strangely in query builder storage explorer version build branch hotfix to master language german russian platform os windows centos macos high sierra architecture regression from not a regression steps to reproduce launch storage explorer open settings application regional settings select deutsche restart storage explorer expand one storage account tables create a new table open query check the localized add new clause expect experience the ‘add new clause’ shows in the same line actual experience the add new clause shows strangely enu screenshot | 0 |
144,032 | 13,093,580,188 | IssuesEvent | 2020-08-03 10:39:56 | alterNERDtive/VoiceAttack-profiles | https://api.github.com/repos/alterNERDtive/VoiceAttack-profiles | closed | Update update instructions | bug documentation done | Apparently importing a profile that already exists (now?) imports it as `<name>-1` instead of overwriting commands. I’ll need to update the update instructions to reflect that:
1. delete profiles
1. import profiles
1. fix includes
:-/ | 1.0 | Update update instructions - Apparently importing a profile that already exists (now?) imports it as `<name>-1` instead of overwriting commands. I’ll need to update the update instructions to reflect that:
1. delete profiles
1. import profiles
1. fix includes
:-/ | non_defect | update update instructions apparently importing a profile that already exists now imports it as instead of overwriting commands i’ll need to update the update instructions to reflect that delete profiles import profiles fix includes | 0 |
51,550 | 7,714,761,653 | IssuesEvent | 2018-05-23 04:04:59 | bcbi/PredictMD.jl | https://api.github.com/repos/bcbi/PredictMD.jl | closed | Use Documenter and mkdocs for documentation | accepted-suggestion documentation | Use Documenter.jl for documentation.
I'm happy to help setting this up if its something that you're interested on | 1.0 | Use Documenter and mkdocs for documentation - Use Documenter.jl for documentation.
I'm happy to help setting this up if its something that you're interested on | non_defect | use documenter and mkdocs for documentation use documenter jl for documentation i m happy to help setting this up if its something that you re interested on | 0 |
14,524 | 17,620,027,942 | IssuesEvent | 2021-08-18 14:20:14 | ORNL-AMO/AMO-Tools-Desktop | https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop | closed | Calculate Binned Weather Results in Cooling Tower Basin calc | enhancement Process Cooling | For each bin set, run calculation method with input parameter = parameter * binned hours. May contain multiple parameters | 1.0 | Calculate Binned Weather Results in Cooling Tower Basin calc - For each bin set, run calculation method with input parameter = parameter * binned hours. May contain multiple parameters | non_defect | calculate binned weather results in cooling tower basin calc for each bin set run calculation method with input parameter parameter binned hours may contain multiple parameters | 0 |
40,782 | 10,163,049,665 | IssuesEvent | 2019-08-07 08:26:05 | opencaching/opencaching-pl | https://api.github.com/repos/opencaching/opencaching-pl | closed | New Password Input should send to main page after login | Component Core Type Defect | Assume you want to reset your password and fill out the web form with that request.
You receive the password reset email.
You click on the link, the browser takes you to the password reset form.
You enter a new password.
Now login using the top login box.
You will be redirected to the same password reset form page (eg. `http://www.opencaching.us/UserAuthorization/newPasswordInput/andrixnet/...` ) which will now display the following text:
_There was a problem with authorization of the password reset code! (More details: Wrong authorization code!)_
This error message is correct when calling the password reset for with a wrong or missing reset code.
Showing this to the user following the click stream described above however should not show such an error. It is misleading.
Login (successful) from newPasswordInput should go to main page. (using top login form)
| 1.0 | New Password Input should send to main page after login - Assume you want to reset your password and fill out the web form with that request.
You receive the password reset email.
You click on the link, the browser takes you to the password reset form.
You enter a new password.
Now login using the top login box.
You will be redirected to the same password reset form page (eg. `http://www.opencaching.us/UserAuthorization/newPasswordInput/andrixnet/...` ) which will now display the following text:
_There was a problem with authorization of the password reset code! (More details: Wrong authorization code!)_
This error message is correct when calling the password reset for with a wrong or missing reset code.
Showing this to the user following the click stream described above however should not show such an error. It is misleading.
Login (successful) from newPasswordInput should go to main page. (using top login form)
| defect | new password input should send to main page after login assume you want to reset your password and fill out the web form with that request you receive the password reset email you click on the link the browser takes you to the password reset form you enter a new password now login using the top login box you will be redirected to the same password reset form page eg which will now display the following text there was a problem with authorization of the password reset code more details wrong authorization code this error message is correct when calling the password reset for with a wrong or missing reset code showing this to the user following the click stream described above however should not show such an error it is misleading login successful from newpasswordinput should go to main page using top login form | 1 |
58,632 | 24,510,489,028 | IssuesEvent | 2022-10-10 20:52:28 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | [Bug]: AWS sso can not associate user to permission set and account | bug service/identitystore service/ssoadmin | ### Terraform Core Version
1.0.0
### AWS Provider Version
4.34
### Affected Resource(s)
```hcl
// version used in 4.33
data "aws_identitystore_user" "users" {
for_each = toset(local.users[*].name)
identity_store_id = tolist(data.aws_ssoadmin_instances.selected.identity_store_ids)[0]
filter {
attribute_path = "UserName"
attribute_value = each.value
}
}
// with no deprecation given by 4.34
data "aws_identitystore_user" "users" {
for_each = toset(local.users[*].name)
identity_store_id = tolist(data.aws_ssoadmin_instances.selected.identity_store_ids)[0]
alternate_identifier {
unique_attribute {
attribute_path = "UserName"
attribute_value = each.value
}
}
}
resource "aws_ssoadmin_account_assignment" "users" {
for_each = {for val in local.users_privileges: replace("${val.name}-${val.account}-${val.permission}", " ", "_") => val}
instance_arn = tolist(data.aws_ssoadmin_instances.selected.arns)[0]
permission_set_arn = data.aws_ssoadmin_permission_set.list[each.value.permission].arn
principal_id = data.aws_identitystore_user.users[each.value.name].user_id
target_id = local.accounts_by_name[each.value.account]
principal_type = "USER"
target_type = "AWS_ACCOUNT"
}
```
### Expected Behavior
with provider 4.33 it associate the user to a permission set and account
### Actual Behavior
With both version of aws_identitystore_user we have the same error:
```
Error: setting AWS SSO Identity Store User Data Source (936*****83c): external_ids.0.id: '' expected type 'bool', got unconvertible type 'string', value: 'b29******94'
with data.aws_identitystore_user.users["name.surname@example.com"],
on data.tf line 5, in data "aws_identitystore_user" "users":
5: data "aws_identitystore_user" "users" {
```
### Relevant Error/Panic Output Snippet
```shell
Error: setting AWS SSO Identity Store User Data Source (936*****83c): external_ids.0.id: '' expected type 'bool', got unconvertible type 'string', value: 'b29******94'
with data.aws_identitystore_user.users["name.surname@example.com"],
on data.tf line 5, in data "aws_identitystore_user" "users":
5: data "aws_identitystore_user" "users" {
```
### Terraform Configuration Files
```
[Container] 2022/10/07 08:50:50 Running command bash adf-build/helpers/terraform/adf_terraform.sh $TF_STAGE
Terraform v1.0.0
on linux_amd64
Your version of Terraform is out of date! The latest version
is 1.3.2. You can update by downloading from https://www.terraform.io/downloads.html
Terraform stage: plan
List of target regions: eu-west-1
List of target account: ****
Running terraform plan on account **** and region eu-west-1
Initializing the backend...
Successfully configured the backend "s3"! Terraform will automatically
use this backend unless the backend configuration changes.
Initializing provider plugins...
- Finding hashicorp/aws versions matching ">= 4.0.0"...
- Installing hashicorp/aws v4.34.0...
- Installed hashicorp/aws v4.34.0 (signed by HashiCorp)
```
### Steps to Reproduce
- have AWS SSO
- have an external IDP (Azure)
- assegnate user to specific account and permission set
### Debug Output
_No response_
### Panic Output
_No response_
### Important Factoids
_No response_
### References
_No response_
### Would you like to implement a fix?
_No response_ | 2.0 | [Bug]: AWS sso can not associate user to permission set and account - ### Terraform Core Version
1.0.0
### AWS Provider Version
4.34
### Affected Resource(s)
```hcl
// version used in 4.33
data "aws_identitystore_user" "users" {
for_each = toset(local.users[*].name)
identity_store_id = tolist(data.aws_ssoadmin_instances.selected.identity_store_ids)[0]
filter {
attribute_path = "UserName"
attribute_value = each.value
}
}
// with no deprecation given by 4.34
data "aws_identitystore_user" "users" {
for_each = toset(local.users[*].name)
identity_store_id = tolist(data.aws_ssoadmin_instances.selected.identity_store_ids)[0]
alternate_identifier {
unique_attribute {
attribute_path = "UserName"
attribute_value = each.value
}
}
}
resource "aws_ssoadmin_account_assignment" "users" {
for_each = {for val in local.users_privileges: replace("${val.name}-${val.account}-${val.permission}", " ", "_") => val}
instance_arn = tolist(data.aws_ssoadmin_instances.selected.arns)[0]
permission_set_arn = data.aws_ssoadmin_permission_set.list[each.value.permission].arn
principal_id = data.aws_identitystore_user.users[each.value.name].user_id
target_id = local.accounts_by_name[each.value.account]
principal_type = "USER"
target_type = "AWS_ACCOUNT"
}
```
### Expected Behavior
with provider 4.33 it associate the user to a permission set and account
### Actual Behavior
With both version of aws_identitystore_user we have the same error:
```
Error: setting AWS SSO Identity Store User Data Source (936*****83c): external_ids.0.id: '' expected type 'bool', got unconvertible type 'string', value: 'b29******94'
with data.aws_identitystore_user.users["name.surname@example.com"],
on data.tf line 5, in data "aws_identitystore_user" "users":
5: data "aws_identitystore_user" "users" {
```
### Relevant Error/Panic Output Snippet
```shell
Error: setting AWS SSO Identity Store User Data Source (936*****83c): external_ids.0.id: '' expected type 'bool', got unconvertible type 'string', value: 'b29******94'
with data.aws_identitystore_user.users["name.surname@example.com"],
on data.tf line 5, in data "aws_identitystore_user" "users":
5: data "aws_identitystore_user" "users" {
```
### Terraform Configuration Files
```
[Container] 2022/10/07 08:50:50 Running command bash adf-build/helpers/terraform/adf_terraform.sh $TF_STAGE
Terraform v1.0.0
on linux_amd64
Your version of Terraform is out of date! The latest version
is 1.3.2. You can update by downloading from https://www.terraform.io/downloads.html
Terraform stage: plan
List of target regions: eu-west-1
List of target account: ****
Running terraform plan on account **** and region eu-west-1
Initializing the backend...
Successfully configured the backend "s3"! Terraform will automatically
use this backend unless the backend configuration changes.
Initializing provider plugins...
- Finding hashicorp/aws versions matching ">= 4.0.0"...
- Installing hashicorp/aws v4.34.0...
- Installed hashicorp/aws v4.34.0 (signed by HashiCorp)
```
### Steps to Reproduce
- have AWS SSO
- have an external IDP (Azure)
- assegnate user to specific account and permission set
### Debug Output
_No response_
### Panic Output
_No response_
### Important Factoids
_No response_
### References
_No response_
### Would you like to implement a fix?
_No response_ | non_defect | aws sso can not associate user to permission set and account terraform core version aws provider version affected resource s hcl version used in data aws identitystore user users for each toset local users name identity store id tolist data aws ssoadmin instances selected identity store ids filter attribute path username attribute value each value with no deprecation given by data aws identitystore user users for each toset local users name identity store id tolist data aws ssoadmin instances selected identity store ids alternate identifier unique attribute attribute path username attribute value each value resource aws ssoadmin account assignment users for each for val in local users privileges replace val name val account val permission val instance arn tolist data aws ssoadmin instances selected arns permission set arn data aws ssoadmin permission set list arn principal id data aws identitystore user users user id target id local accounts by name principal type user target type aws account expected behavior with provider it associate the user to a permission set and account actual behavior with both version of aws identitystore user we have the same error error setting aws sso identity store user data source external ids id expected type bool got unconvertible type string value with data aws identitystore user users on data tf line in data aws identitystore user users data aws identitystore user users relevant error panic output snippet shell error setting aws sso identity store user data source external ids id expected type bool got unconvertible type string value with data aws identitystore user users on data tf line in data aws identitystore user users data aws identitystore user users terraform configuration files running command bash adf build helpers terraform adf terraform sh tf stage terraform on linux your version of terraform is out of date the latest version is you can update by downloading from terraform stage plan list of target regions eu west list of target account running terraform plan on account and region eu west initializing the backend successfully configured the backend terraform will automatically use this backend unless the backend configuration changes initializing provider plugins finding hashicorp aws versions matching installing hashicorp aws installed hashicorp aws signed by hashicorp steps to reproduce have aws sso have an external idp azure assegnate user to specific account and permission set debug output no response panic output no response important factoids no response references no response would you like to implement a fix no response | 0 |
294,862 | 9,049,568,984 | IssuesEvent | 2019-02-12 05:20:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.google.com - see bug description | browser-firefox-mobile priority-critical | <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=madrocks.restaurant&client=firefox-b-1-m&biw=360&bih=560&tbm=isch&oq=madrocks.restaurant&gs_l=mobile-heirloom-serp.3...14403.14518.0.14608.11.2.0.0.0.0.0.0..0.0....0...1c.1j4.34.mobile-heirloom-serp..11.0.0.-dA5k1Fr39k#mhpiv=0&spf=1549946082003
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: wrong website listing for business
**Steps to Reproduce**:
old website - new website is https://madrocks.restaurant - former website designer is holding it hostage and wont delete it because he is mad he doesnt manage it anymore and a new company does.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190211092917</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.google.com - see bug description - <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=madrocks.restaurant&client=firefox-b-1-m&biw=360&bih=560&tbm=isch&oq=madrocks.restaurant&gs_l=mobile-heirloom-serp.3...14403.14518.0.14608.11.2.0.0.0.0.0.0..0.0....0...1c.1j4.34.mobile-heirloom-serp..11.0.0.-dA5k1Fr39k#mhpiv=0&spf=1549946082003
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: wrong website listing for business
**Steps to Reproduce**:
old website - new website is https://madrocks.restaurant - former website designer is holding it hostage and wont delete it because he is mad he doesnt manage it anymore and a new company does.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190211092917</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_defect | see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description wrong website listing for business steps to reproduce old website new website is former website designer is holding it hostage and wont delete it because he is mad he doesnt manage it anymore and a new company does browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel nightly from with ❤️ | 0 |
47,231 | 7,311,490,562 | IssuesEvent | 2018-02-28 17:55:19 | kuzzleio/kuzzle-sdk | https://api.github.com/repos/kuzzleio/kuzzle-sdk | closed | Give a minimal example of how to instantiate a Kuzzle object in README | need-documentation priority-normal sdk-js | In the JS SDK, there is no example on how to instantiate a Kuzzle object. | 1.0 | Give a minimal example of how to instantiate a Kuzzle object in README - In the JS SDK, there is no example on how to instantiate a Kuzzle object. | non_defect | give a minimal example of how to instantiate a kuzzle object in readme in the js sdk there is no example on how to instantiate a kuzzle object | 0 |
75,687 | 9,881,871,662 | IssuesEvent | 2019-06-24 15:33:16 | apauner/Pytures | https://api.github.com/repos/apauner/Pytures | opened | DR-0001: Main window design | documentation | _Info: this section contains the stuff that will appear when the user calls the aplication._
- _Menu._
- _Project tree._
- _Graphic window._ | 1.0 | DR-0001: Main window design - _Info: this section contains the stuff that will appear when the user calls the aplication._
- _Menu._
- _Project tree._
- _Graphic window._ | non_defect | dr main window design info this section contains the stuff that will appear when the user calls the aplication menu project tree graphic window | 0 |
217,852 | 7,328,504,284 | IssuesEvent | 2018-03-04 21:20:03 | ameliejyc/prereq-check | https://api.github.com/repos/ameliejyc/prereq-check | closed | Team member login | bug priority-1 | As tested with Dan, unable to view accounts.
Error occurs when on heroku app used on chromebook (not mac).
Works when run locally on mac. | 1.0 | Team member login - As tested with Dan, unable to view accounts.
Error occurs when on heroku app used on chromebook (not mac).
Works when run locally on mac. | non_defect | team member login as tested with dan unable to view accounts error occurs when on heroku app used on chromebook not mac works when run locally on mac | 0 |
168,212 | 6,366,444,980 | IssuesEvent | 2017-08-01 01:32:11 | EyeSeeTea/SurveillanceLaosApp | https://api.github.com/repos/EyeSeeTea/SurveillanceLaosApp | closed | Invalid result layout | complexity - med (1-5hr) priority - high type - cosmetic (layout) | For Prompt 2 Invalid Test Result they want to change the artwork to make it more clear Orange Button 1 (repeat RDT test, count the RDT stock) vs. Orange Button 2 (go back, don't count the RDT stock).
I think this changes the layout. Can you look?
Graphic: https://drive.google.com/open?id=0BzWNKmbLqoIRQ1dUaUNXSDk1V1k
Translation (Prompt 2): https://drive.google.com/drive/u/1/folders/0BzWNKmbLqoIRenFHMUxnckhlV2c
| 1.0 | Invalid result layout - For Prompt 2 Invalid Test Result they want to change the artwork to make it more clear Orange Button 1 (repeat RDT test, count the RDT stock) vs. Orange Button 2 (go back, don't count the RDT stock).
I think this changes the layout. Can you look?
Graphic: https://drive.google.com/open?id=0BzWNKmbLqoIRQ1dUaUNXSDk1V1k
Translation (Prompt 2): https://drive.google.com/drive/u/1/folders/0BzWNKmbLqoIRenFHMUxnckhlV2c
| non_defect | invalid result layout for prompt invalid test result they want to change the artwork to make it more clear orange button repeat rdt test count the rdt stock vs orange button go back don t count the rdt stock i think this changes the layout can you look graphic translation prompt | 0 |
758,887 | 26,572,627,317 | IssuesEvent | 2023-01-21 11:21:43 | pibolib/hack16-2 | https://api.github.com/repos/pibolib/hack16-2 | closed | Adjust player model to account for new changes | visuals priority:med | Specifications:
Add more movement in response to player movement to body parts.
- Legs should lag behind player
- Hair sways according to movement and lags behind player slightly
- Body parts recoil slightly in response to player gunfire
- Boots->Hovering element should emit particles
- Gun should emit particles when firing | 1.0 | Adjust player model to account for new changes - Specifications:
Add more movement in response to player movement to body parts.
- Legs should lag behind player
- Hair sways according to movement and lags behind player slightly
- Body parts recoil slightly in response to player gunfire
- Boots->Hovering element should emit particles
- Gun should emit particles when firing | non_defect | adjust player model to account for new changes specifications add more movement in response to player movement to body parts legs should lag behind player hair sways according to movement and lags behind player slightly body parts recoil slightly in response to player gunfire boots hovering element should emit particles gun should emit particles when firing | 0 |
33,627 | 14,162,837,854 | IssuesEvent | 2020-11-12 00:49:00 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | opened | Hide ViaPartitionKey property from ServiceBusMessage/ServiceBusReceivedMessage | Service Bus | We can expose it and come up with the name for it once we add the cross entity transaction support.
| 1.0 | Hide ViaPartitionKey property from ServiceBusMessage/ServiceBusReceivedMessage - We can expose it and come up with the name for it once we add the cross entity transaction support.
| non_defect | hide viapartitionkey property from servicebusmessage servicebusreceivedmessage we can expose it and come up with the name for it once we add the cross entity transaction support | 0 |
70,176 | 9,380,522,274 | IssuesEvent | 2019-04-04 17:18:04 | 19majkel94/type-graphql | https://api.github.com/repos/19majkel94/type-graphql | closed | Adding instructions for Typescript/Babel 7 setup | Community :family_man_girl: Documentation :book: Help Wanted :sos: | Currently this framework does not work with Babel 7 and `@babel/preset-typescript` setup, the reasons are:
1- Parameter decorators is not supported by `@babel/plugin-proposal-decorators`
2- Babel strips type information thus `Reflect.getMetadata()` returns undefined, more explanation [url](https://stackoverflow.com/a/54360611)
After spending few hours to come to this conclusion i solved these issues by installing this babel plugin `babel-plugin-transform-typescript-metadata` [url](https://github.com/leonardfactory/babel-plugin-transform-typescript-metadata)
I liked this framework and i almost gave up on it because of these issues,
It will be good to include this in the installation guide, to make it easy for developers with Babel 7 and `@babel/preset-typescript` setup which is a common setup.
Thanks
| 1.0 | Adding instructions for Typescript/Babel 7 setup - Currently this framework does not work with Babel 7 and `@babel/preset-typescript` setup, the reasons are:
1- Parameter decorators is not supported by `@babel/plugin-proposal-decorators`
2- Babel strips type information thus `Reflect.getMetadata()` returns undefined, more explanation [url](https://stackoverflow.com/a/54360611)
After spending few hours to come to this conclusion i solved these issues by installing this babel plugin `babel-plugin-transform-typescript-metadata` [url](https://github.com/leonardfactory/babel-plugin-transform-typescript-metadata)
I liked this framework and i almost gave up on it because of these issues,
It will be good to include this in the installation guide, to make it easy for developers with Babel 7 and `@babel/preset-typescript` setup which is a common setup.
Thanks
| non_defect | adding instructions for typescript babel setup currently this framework does not work with babel and babel preset typescript setup the reasons are parameter decorators is not supported by babel plugin proposal decorators babel strips type information thus reflect getmetadata returns undefined more explanation after spending few hours to come to this conclusion i solved these issues by installing this babel plugin babel plugin transform typescript metadata i liked this framework and i almost gave up on it because of these issues it will be good to include this in the installation guide to make it easy for developers with babel and babel preset typescript setup which is a common setup thanks | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.