Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
235,572 | 25,955,212,940 | IssuesEvent | 2022-12-18 05:33:53 | Dima2022/JS-Demo | https://api.github.com/repos/Dima2022/JS-Demo | closed | CVE-2021-37712 (High) detected in tar-2.2.1.tgz - autoclosed | security vulnerability | ## CVE-2021-37712 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37712>CVE-2021-37712</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: tar - 4.4.18,5.0.10,6.1.9</p>
</p>
</details>
<p></p>
| True | CVE-2021-37712 (High) detected in tar-2.2.1.tgz - autoclosed - ## CVE-2021-37712 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- :x: **tar-2.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37712>CVE-2021-37712</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: tar - 4.4.18,5.0.10,6.1.9</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in tar tgz autoclosed cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file package json path to vulnerable library node modules npm node modules tar package json dependency hierarchy grunt npm install tgz root library npm tgz x tar tgz vulnerable library found in head commit a href found in base branch master vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value additionally on windows systems long path portions would resolve to the same file system entities as their short path counterparts a specially crafted tar archive could thus include a directory with one form of the path followed by a symbolic link with a different string that resolves to the same file system entity followed by a file using the first form by first creating a directory and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar | 0 |
97,989 | 20,597,255,802 | IssuesEvent | 2022-03-05 17:57:00 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.1]Saving update key into extra_query in #__update_sites leaves update site checked out | No Code Attached Yet | ### Steps to reproduce the issue
Install a third party extension that requires a download key like Akeeba Backup Pro
Go to System - Update - Update Sites.
The installed third party extension displays the text 'The Download Key key missing' in red background.
Select the update site.
Enter the required value into the field Download Key.
Save and Close.
The extension shows the text Download Key in blue background and the value of the download key.
Go to System - Maintenance.
### Expected result
Global Check-in shows no items to be checked in.
### Actual result
Global Check-in shows that an item needs to be checked in.
### System information (as much as possible)
Joomla 4.1.1-dev of 5 March 2022, Wampserver 3.2.7 with PHP 8.0.15
### Additional comments
In each case the row in #__update_sites has the following values:
- checked_out 0
- checked_out_time 0000-00-00 00:00:00
The issue was reproduced in 4.1.0 and 4.1.1-dev with the latest versions of Akeeba Backup Pro, Akeeba Admin Tools Pro and JCE Pro.
 | 1.0 | [4.1]Saving update key into extra_query in #__update_sites leaves update site checked out - ### Steps to reproduce the issue
Install a third party extension that requires a download key like Akeeba Backup Pro
Go to System - Update - Update Sites.
The installed third party extension displays the text 'The Download Key key missing' in red background.
Select the update site.
Enter the required value into the field Download Key.
Save and Close.
The extension shows the text Download Key in blue background and the value of the download key.
Go to System - Maintenance.
### Expected result
Global Check-in shows no items to be checked in.
### Actual result
Global Check-in shows that an item needs to be checked in.
### System information (as much as possible)
Joomla 4.1.1-dev of 5 March 2022, Wampserver 3.2.7 with PHP 8.0.15
### Additional comments
In each case the row in #__update_sites has the following values:
- checked_out 0
- checked_out_time 0000-00-00 00:00:00
The issue was reproduced in 4.1.0 and 4.1.1-dev with the latest versions of Akeeba Backup Pro, Akeeba Admin Tools Pro and JCE Pro.
 | non_priority | saving update key into extra query in update sites leaves update site checked out steps to reproduce the issue install a third party extension that requires a download key like akeeba backup pro go to system update update sites the installed third party extension displays the text the download key key missing in red background select the update site enter the required value into the field download key save and close the extension shows the text download key in blue background and the value of the download key go to system maintenance expected result global check in shows no items to be checked in actual result global check in shows that an item needs to be checked in system information as much as possible joomla dev of march wampserver with php additional comments in each case the row in update sites has the following values checked out checked out time the issue was reproduced in and dev with the latest versions of akeeba backup pro akeeba admin tools pro and jce pro | 0 |
363,674 | 25,461,261,945 | IssuesEvent | 2022-11-24 19:26:14 | runfinch/finch | https://api.github.com/repos/runfinch/finch | closed | The logo isn't visible in the README with the dark theme | bug documentation | **Describe the bug**
Briefly describe the problem you are having.
The logo isn't visible in the README with the dark theme
**Steps to reproduce**
A clear, step-by-step set of instructions to reproduce the bug.
Enable dark theme on MacOS, view the readme with Safari
**Expected behavior**
Description of what you expected to happen.
See the logo
Initially, I thought there was no readme until I scrolled down.
**Screenshots or logs**
If applicable, add screenshots or logs to help explain your problem.
<img width="1426" alt="Screenshot 2022-11-23 at 08 46 13" src="https://user-images.githubusercontent.com/6358735/203504158-733c2524-84dc-4c9f-930a-4d955f1ec926.png">
| 1.0 | The logo isn't visible in the README with the dark theme - **Describe the bug**
Briefly describe the problem you are having.
The logo isn't visible in the README with the dark theme
**Steps to reproduce**
A clear, step-by-step set of instructions to reproduce the bug.
Enable dark theme on MacOS, view the readme with Safari
**Expected behavior**
Description of what you expected to happen.
See the logo
Initially, I thought there was no readme until I scrolled down.
**Screenshots or logs**
If applicable, add screenshots or logs to help explain your problem.
<img width="1426" alt="Screenshot 2022-11-23 at 08 46 13" src="https://user-images.githubusercontent.com/6358735/203504158-733c2524-84dc-4c9f-930a-4d955f1ec926.png">
| non_priority | the logo isn t visible in the readme with the dark theme describe the bug briefly describe the problem you are having the logo isn t visible in the readme with the dark theme steps to reproduce a clear step by step set of instructions to reproduce the bug enable dark theme on macos view the readme with safari expected behavior description of what you expected to happen see the logo initially i thought there was no readme until i scrolled down screenshots or logs if applicable add screenshots or logs to help explain your problem img width alt screenshot at src | 0 |
19,600 | 10,470,980,633 | IssuesEvent | 2019-09-23 06:25:29 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | readLine is slow (and doesn't match Python's definition of a "line") | Performance Stdlib | Perhaps this should be 3 GitHub issues? But maybe fixing one problem will fix the other.
# Summary
1. `streams.readLine` is slow
2. `streams.readLine` chokes on null bytes
3. `streams.readLine` doesn't produce the same "lines" as either Python or `wc` (this or more surprising than it is a "bug")
# Details
I'm porting a Node project to Nim and noticed that the Nim version is much slower (about twice as slow) as the Node version. So I ran some comparisons [here](https://gist.github.com/iffy/ede3fb5b637b34566ecd36c447995c75).
To test what's going on, I made 4 scripts that read through each line of a file and print out the number of lines found.
1. Node
2. Nim (using builtin `streams.readLine`)
3. Nim (using custom-made `readLine` proc)
4. Python
Each script was run against two files:
1. Mostly binary data (with many newlines)
2. Only string data
Here is my custom-made `readLine`:
```nim
proc readLine(s: Stream, line: var TaintedString, delimiter = '\n'): bool =
line.string.setLen(0)
result = true
while result:
var c = readChar(s)
if c == delimiter:
break
elif c == '\0':
if s.atEnd():
result = false
line.string.add(c)
```
## Accuracy Results
1. `wc`, Python and my custom-made `readLine` proc all agree on what a "line" is (i.e. they print out the same number for "number of lines")
2. Node reports more lines than `wc`
3. `streams.readLine` reports more lines than `wc` on text-only files (probably because it counts `\L` too)
4. `streams.readLine` chokes if there are null bytes
## Speed Results
| program | inputfile | time | line count |
|---|---|---|---|
| wc | hugebin.txt | | 1444473 |
| reader.js | hugebin.txt | 2.18 | 1844188 |
| Nim (custom proc) | hugebin.txt | 5.13 | 1444473 |
| `streams.readLine` | hugebin.txt | 0 | 279 |
| reader.py | hugebin.txt | 0.5 | 1444473 |
| wc | hugestr.txt | | 2078017 |
| reader.js | hugestr.txt | 0.65 | 3095841 |
| Nim (custom proc) | hugestr.txt | 5.11 | 2078017 |
| `streams.readLine` | hugestr.txt | 5.01 | 3085918 |
| reader.py | hugestr.txt | 0.64 | 2078017 |
| True | readLine is slow (and doesn't match Python's definition of a "line") - Perhaps this should be 3 GitHub issues? But maybe fixing one problem will fix the other.
# Summary
1. `streams.readLine` is slow
2. `streams.readLine` chokes on null bytes
3. `streams.readLine` doesn't produce the same "lines" as either Python or `wc` (this or more surprising than it is a "bug")
# Details
I'm porting a Node project to Nim and noticed that the Nim version is much slower (about twice as slow) as the Node version. So I ran some comparisons [here](https://gist.github.com/iffy/ede3fb5b637b34566ecd36c447995c75).
To test what's going on, I made 4 scripts that read through each line of a file and print out the number of lines found.
1. Node
2. Nim (using builtin `streams.readLine`)
3. Nim (using custom-made `readLine` proc)
4. Python
Each script was run against two files:
1. Mostly binary data (with many newlines)
2. Only string data
Here is my custom-made `readLine`:
```nim
proc readLine(s: Stream, line: var TaintedString, delimiter = '\n'): bool =
line.string.setLen(0)
result = true
while result:
var c = readChar(s)
if c == delimiter:
break
elif c == '\0':
if s.atEnd():
result = false
line.string.add(c)
```
## Accuracy Results
1. `wc`, Python and my custom-made `readLine` proc all agree on what a "line" is (i.e. they print out the same number for "number of lines")
2. Node reports more lines than `wc`
3. `streams.readLine` reports more lines than `wc` on text-only files (probably because it counts `\L` too)
4. `streams.readLine` chokes if there are null bytes
## Speed Results
| program | inputfile | time | line count |
|---|---|---|---|
| wc | hugebin.txt | | 1444473 |
| reader.js | hugebin.txt | 2.18 | 1844188 |
| Nim (custom proc) | hugebin.txt | 5.13 | 1444473 |
| `streams.readLine` | hugebin.txt | 0 | 279 |
| reader.py | hugebin.txt | 0.5 | 1444473 |
| wc | hugestr.txt | | 2078017 |
| reader.js | hugestr.txt | 0.65 | 3095841 |
| Nim (custom proc) | hugestr.txt | 5.11 | 2078017 |
| `streams.readLine` | hugestr.txt | 5.01 | 3085918 |
| reader.py | hugestr.txt | 0.64 | 2078017 |
| non_priority | readline is slow and doesn t match python s definition of a line perhaps this should be github issues but maybe fixing one problem will fix the other summary streams readline is slow streams readline chokes on null bytes streams readline doesn t produce the same lines as either python or wc this or more surprising than it is a bug details i m porting a node project to nim and noticed that the nim version is much slower about twice as slow as the node version so i ran some comparisons to test what s going on i made scripts that read through each line of a file and print out the number of lines found node nim using builtin streams readline nim using custom made readline proc python each script was run against two files mostly binary data with many newlines only string data here is my custom made readline nim proc readline s stream line var taintedstring delimiter n bool line string setlen result true while result var c readchar s if c delimiter break elif c if s atend result false line string add c accuracy results wc python and my custom made readline proc all agree on what a line is i e they print out the same number for number of lines node reports more lines than wc streams readline reports more lines than wc on text only files probably because it counts l too streams readline chokes if there are null bytes speed results program inputfile time line count wc hugebin txt reader js hugebin txt nim custom proc hugebin txt streams readline hugebin txt reader py hugebin txt wc hugestr txt reader js hugestr txt nim custom proc hugestr txt streams readline hugestr txt reader py hugestr txt | 0 |
50,444 | 12,511,775,679 | IssuesEvent | 2020-06-02 21:14:05 | microsoft/vscode-cpptools | https://api.github.com/repos/microsoft/vscode-cpptools | closed | "preLaunchTask" in launch.json and "label" in task.json are different while generating them by default | Language Service build/tasks fixed (release pending) quick fix | The default generated launch.json and task.json files are below:
>I don't know why there are some Chinese words. My VSCode is in English and doesn't install the language package. I think it's an another bug.
```
{
// Use IntelliSense to learn about possible attributes.
// Hover to view descriptions of existing attributes.
// For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
"version": "0.2.0",
"configurations": [
{
"name": "gcc.exe - 生成和调试活动文件",
"type": "cppdbg",
"request": "launch",
"program": "${fileDirname}\\${fileBasenameNoExtension}.exe",
"args": [],
"stopAtEntry": false,
"cwd": "${workspaceFolder}",
"environment": [],
"externalConsole": false,
"MIMode": "gdb",
"miDebuggerPath": "C:\\ProgramData\\chocolatey\\bin\\gdb.exe",
"setupCommands": [
{
"description": "为 gdb 启用整齐打印",
"text": "-enable-pretty-printing",
"ignoreFailures": true
}
],
"preLaunchTask": "gcc.exe build active file"
}
]
}
```
```
{
"version": "2.0.0",
"tasks": [
{
"type": "shell",
"label": "shell: gcc.exe build active file",
"command": "C:\\ProgramData\\chocolatey\\bin\\gcc.exe",
"args": [
"-g",
"${file}",
"-o",
"${fileDirname}\\${fileBasenameNoExtension}.exe"
],
"options": {
"cwd": "C:\\ProgramData\\chocolatey\\bin"
},
"problemMatcher": [
"$gcc"
],
"group": "build"
}
]
}
```
Notice the different:
```
"preLaunchTask": "gcc.exe build active file" //from the launch.json
"label": "shell: gcc.exe build active file" //from the task.json
```
They are not the same. It will cause an error when press F5 on the keyboard.

If correct them as the same, the bug gone.
```
"preLaunchTask": "gcc.exe build active file" //from the launch.json
"label": "gcc.exe build active file" //from the task.json
``` | 1.0 | "preLaunchTask" in launch.json and "label" in task.json are different while generating them by default - The default generated launch.json and task.json files are below:
>I don't know why there are some Chinese words. My VSCode is in English and doesn't install the language package. I think it's an another bug.
```
{
// Use IntelliSense to learn about possible attributes.
// Hover to view descriptions of existing attributes.
// For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
"version": "0.2.0",
"configurations": [
{
"name": "gcc.exe - 生成和调试活动文件",
"type": "cppdbg",
"request": "launch",
"program": "${fileDirname}\\${fileBasenameNoExtension}.exe",
"args": [],
"stopAtEntry": false,
"cwd": "${workspaceFolder}",
"environment": [],
"externalConsole": false,
"MIMode": "gdb",
"miDebuggerPath": "C:\\ProgramData\\chocolatey\\bin\\gdb.exe",
"setupCommands": [
{
"description": "为 gdb 启用整齐打印",
"text": "-enable-pretty-printing",
"ignoreFailures": true
}
],
"preLaunchTask": "gcc.exe build active file"
}
]
}
```
```
{
"version": "2.0.0",
"tasks": [
{
"type": "shell",
"label": "shell: gcc.exe build active file",
"command": "C:\\ProgramData\\chocolatey\\bin\\gcc.exe",
"args": [
"-g",
"${file}",
"-o",
"${fileDirname}\\${fileBasenameNoExtension}.exe"
],
"options": {
"cwd": "C:\\ProgramData\\chocolatey\\bin"
},
"problemMatcher": [
"$gcc"
],
"group": "build"
}
]
}
```
Notice the different:
```
"preLaunchTask": "gcc.exe build active file" //from the launch.json
"label": "shell: gcc.exe build active file" //from the task.json
```
They are not the same. It will cause an error when press F5 on the keyboard.

If correct them as the same, the bug gone.
```
"preLaunchTask": "gcc.exe build active file" //from the launch.json
"label": "gcc.exe build active file" //from the task.json
``` | non_priority | prelaunchtask in launch json and label in task json are different while generating them by default the default generated launch json and task json files are below i don t know why there are some chinese words my vscode is in english and doesn t install the language package i think it s an another bug use intellisense to learn about possible attributes hover to view descriptions of existing attributes for more information visit version configurations name gcc exe 生成和调试活动文件 type cppdbg request launch program filedirname filebasenamenoextension exe args stopatentry false cwd workspacefolder environment externalconsole false mimode gdb midebuggerpath c programdata chocolatey bin gdb exe setupcommands description 为 gdb 启用整齐打印 text enable pretty printing ignorefailures true prelaunchtask gcc exe build active file version tasks type shell label shell gcc exe build active file command c programdata chocolatey bin gcc exe args g file o filedirname filebasenamenoextension exe options cwd c programdata chocolatey bin problemmatcher gcc group build notice the different prelaunchtask gcc exe build active file from the launch json label shell gcc exe build active file from the task json they are not the same it will cause an error when press on the keyboard if correct them as the same the bug gone prelaunchtask gcc exe build active file from the launch json label gcc exe build active file from the task json | 0 |
132,747 | 28,314,536,137 | IssuesEvent | 2023-04-10 18:22:04 | Tuxemon/Tuxemon | https://api.github.com/repos/Tuxemon/Tuxemon | closed | Variable_math turned my integer into a float | Code Quality Event | **Describe the bug**
I'm using the game variable 'money' to hold player's money.
I used the following 2 commands:
```
set_variable money:0
variable_math money,+,100,money
```
It incremented money by 100, but also turned it into a float - now it's 100.0
Not a huge deal, but now my variable is a float? Would that be annoying in the future? Would
``` is variable_set money:100 ```
fail? I guess I can just ignore it for now, so it's low/no priority. But I wanted to mention it. | 1.0 | Variable_math turned my integer into a float - **Describe the bug**
I'm using the game variable 'money' to hold player's money.
I used the following 2 commands:
```
set_variable money:0
variable_math money,+,100,money
```
It incremented money by 100, but also turned it into a float - now it's 100.0
Not a huge deal, but now my variable is a float? Would that be annoying in the future? Would
``` is variable_set money:100 ```
fail? I guess I can just ignore it for now, so it's low/no priority. But I wanted to mention it. | non_priority | variable math turned my integer into a float describe the bug i m using the game variable money to hold player s money i used the following commands set variable money variable math money money it incremented money by but also turned it into a float now it s not a huge deal but now my variable is a float would that be annoying in the future would is variable set money fail i guess i can just ignore it for now so it s low no priority but i wanted to mention it | 0 |
34,999 | 30,668,374,056 | IssuesEvent | 2023-07-25 20:10:11 | w3c/aria-practices | https://api.github.com/repos/w3c/aria-practices | closed | Infrastructure: Remove temp pinning to Node 16 to work around tests failing with Node 18 in CI | bug Infrastructure regression-testing | Initially reported by @jongund through an email thread and further supported by the latest set of [dependabot updates](https://github.com/w3c/aria-practices/pulls?q=is%3Apr+is%3Aopen+dependabot), all the regression tests are failing with a similar error for each test, when ran with Node 18 in the GitHub Actions:
```
not ok 1 - accordion_accordion › before hook
---
name: AssertionError
message: Rejected promise returned by test
values:
'Rejected promise returned by test. Reason:': |-
Error {
message: 'Timed out while waiting for WebDriver server',
}
at: 'Timeout.poll [as _onTimeout] (test/util/start-geckodriver.js:34:16)'
...
```
The error message is `Timed out while waiting for WebDriver server`. | 1.0 | Infrastructure: Remove temp pinning to Node 16 to work around tests failing with Node 18 in CI - Initially reported by @jongund through an email thread and further supported by the latest set of [dependabot updates](https://github.com/w3c/aria-practices/pulls?q=is%3Apr+is%3Aopen+dependabot), all the regression tests are failing with a similar error for each test, when ran with Node 18 in the GitHub Actions:
```
not ok 1 - accordion_accordion › before hook
---
name: AssertionError
message: Rejected promise returned by test
values:
'Rejected promise returned by test. Reason:': |-
Error {
message: 'Timed out while waiting for WebDriver server',
}
at: 'Timeout.poll [as _onTimeout] (test/util/start-geckodriver.js:34:16)'
...
```
The error message is `Timed out while waiting for WebDriver server`. | non_priority | infrastructure remove temp pinning to node to work around tests failing with node in ci initially reported by jongund through an email thread and further supported by the latest set of all the regression tests are failing with a similar error for each test when ran with node in the github actions not ok accordion accordion › before hook name assertionerror message rejected promise returned by test values rejected promise returned by test reason error message timed out while waiting for webdriver server at timeout poll test util start geckodriver js the error message is timed out while waiting for webdriver server | 0 |
254,400 | 19,212,628,024 | IssuesEvent | 2021-12-07 04:59:29 | NorthernSantan/Rectangle-Program | https://api.github.com/repos/NorthernSantan/Rectangle-Program | opened | Quit Option Needs Update | documentation enhancement Unimplemented | The quit option needs to do the following.

We do need to print out the text into the file using the format from option 6.
Also, read me needs to be updated. FileWriter is NOT deprecated; its actual usage isn't implemented. | 1.0 | Quit Option Needs Update - The quit option needs to do the following.

We do need to print out the text into the file using the format from option 6.
Also, read me needs to be updated. FileWriter is NOT deprecated; its actual usage isn't implemented. | non_priority | quit option needs update the quit option needs to do the following we do need to print out the text into the file using the format from option also read me needs to be updated filewriter is not deprecated its actual usage isn t implemented | 0 |
42,013 | 5,414,616,856 | IssuesEvent | 2017-03-01 19:31:36 | AffiliateWP/AffiliateWP | https://api.github.com/repos/AffiliateWP/AffiliateWP | closed | Edit referral amount on existing referral disassociates it from the affiliate. | bug Unit Tests | Beta bug report form this ticket - I confirmed it on my test site using 2.0 released version.
https://secure.helpscout.net/conversation/309619819/51624/?folderId=205529
Steps to recreate:
Edit an existing referral amount, click to save changes and it appears to get separated from the affiliate. The status for my referral was Unpaid but not sure that is pertinent.
Screen shot from the user:

| 1.0 | Edit referral amount on existing referral disassociates it from the affiliate. - Beta bug report form this ticket - I confirmed it on my test site using 2.0 released version.
https://secure.helpscout.net/conversation/309619819/51624/?folderId=205529
Steps to recreate:
Edit an existing referral amount, click to save changes and it appears to get separated from the affiliate. The status for my referral was Unpaid but not sure that is pertinent.
Screen shot from the user:

| non_priority | edit referral amount on existing referral disassociates it from the affiliate beta bug report form this ticket i confirmed it on my test site using released version steps to recreate edit an existing referral amount click to save changes and it appears to get separated from the affiliate the status for my referral was unpaid but not sure that is pertinent screen shot from the user | 0 |
66,313 | 8,914,959,671 | IssuesEvent | 2019-01-19 00:46:37 | PegaSysEng/pantheon | https://api.github.com/repos/PegaSysEng/pantheon | closed | Add eth_estimateGas use cases | documentation | Add use cases for:
- Estimating how much gas to set when making a transaction to ensure we set enough for the transfer to complete
- Showing the cost of an action on a dApp before the user makes a transaction. | 1.0 | Add eth_estimateGas use cases - Add use cases for:
- Estimating how much gas to set when making a transaction to ensure we set enough for the transfer to complete
- Showing the cost of an action on a dApp before the user makes a transaction. | non_priority | add eth estimategas use cases add use cases for estimating how much gas to set when making a transaction to ensure we set enough for the transfer to complete showing the cost of an action on a dapp before the user makes a transaction | 0 |
343,934 | 24,790,974,090 | IssuesEvent | 2022-10-24 13:44:38 | alexjustesen/speedtest-tracker | https://api.github.com/repos/alexjustesen/speedtest-tracker | closed | Improve README | 📖 documentation | When opening project's README it would be beneficial to have some explanation what this project about for those who has no idea. Would be nice, if you could add a screenshot (or couple of them, if there are more than one view) and feature list.
| 1.0 | Improve README - When opening project's README it would be beneficial to have some explanation what this project about for those who has no idea. Would be nice, if you could add a screenshot (or couple of them, if there are more than one view) and feature list.
| non_priority | improve readme when opening project s readme it would be beneficial to have some explanation what this project about for those who has no idea would be nice if you could add a screenshot or couple of them if there are more than one view and feature list | 0 |
60,060 | 6,670,315,443 | IssuesEvent | 2017-10-03 22:56:57 | equella/Equella | https://api.github.com/repos/equella/Equella | closed | Server administration settings page: still has reference to licences | bug Ready for 6.5 GA Testing | This text should be either removed or replaced with something more appropriate. I think system notifications are only used for filestore usage warnings

| 1.0 | Server administration settings page: still has reference to licences - This text should be either removed or replaced with something more appropriate. I think system notifications are only used for filestore usage warnings

| non_priority | server administration settings page still has reference to licences this text should be either removed or replaced with something more appropriate i think system notifications are only used for filestore usage warnings | 0 |
99 | 3,835,669,237 | IssuesEvent | 2016-04-01 15:08:15 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | opened | SIGABRT_System.Globalization.Native.so!UErrorCodeToBool | bug reliability | **The notes in this bug are referencing the dump [projectk-23923-00-xlsh8-amd64chk_01A2.zip](https://dotnetjobresults.blob.core.windows.net/projectk-23923-00-xlsh8-stress/f7b98499-069d-4c64-b420-af5c211233a3/stress/generated/00_01hr/projectk-23923-00-xlsh8-amd64chk_01A2/projectk-23923-00-xlsh8-amd64chk_01A2.csproj/projectk-23923-00-xlsh8-amd64chk_01A2.zip?sv=2015-04-05&sr=c&sig=O%2FEO8ALizJpZh%2Bmm6lQNB%2BPinsMWLz17JO24BdFDg4U%3D&se=2017-03-29T07:34:00Z&sp=rw), more dumps are availabie for this issue if needed**
**This failure is due to an assert in System.Globalization.Native, because a call to GetLocal returned a status of U_BUFFER_OVERFLOW_ERROR. However there seems to be more going on in this dump, as the stack of this thread seems to be torn either with a corrupted stack or corrupted unwind info.**
STOP_REASON:
SIGABRT
FAILURE_HASH:
SIGABRT_System.Globalization.Native.so!UErrorCodeToBool
FAULT_STACK:
libc.so.6!__GI_raise
libc.so.6!__GI_abort
libc.so.6!__assert_fail_base
libc.so.6!__GI___assert_fail
System.Globalization.Native.so!UErrorCodeToBool
System.Globalization.Native.so!GlobalizationNative_GetLocaleName
mscorlib.ni!DomainNeutralILStubClass.IL_STUB_PInvoke
UNKNOWN!UNKNOWN
FAULT_SYMBOL:
System.Globalization.Native.so!UErrorCodeToBool
**From looking at the values from frame 5 and the code we can see that we are asserting because a status of U_BUFFER_OVERFLOW_ERROR was returned from GetLocal**
(lldb) fr s 5
frame #5: 0x00007f68dd5d4302 System.Globalization.Native.so`::GlobalizationNative_GetLocaleName(localeName=0x00007f68480300f4, value=0x00007f68d8d68a60, valueLength=157) + 210 at locale.cpp:130
(lldb) fr v
(const UChar *) localeName = 0x00007f68480300f4
(int32_t) valueLength = 157
(UChar *) value = 0x00007f68d8d68a60
(UErrorCode) status = U_BUFFER_OVERFLOW_ERROR
(char [157]) localeNameBuffer = "notavalidculture"
extern "C" int32_t GlobalizationNative_GetLocaleName(const UChar* localeName, UChar* value, int32_t valueLength)
{
UErrorCode status = U_ZERO_ERROR;
char localeNameBuffer[ULOC_FULLNAME_CAPACITY];
GetLocale(localeName, localeNameBuffer, ULOC_FULLNAME_CAPACITY, true, &status);
...
return UErrorCodeToBool(status);
}
int32_t UErrorCodeToBool(UErrorCode status)
{
...
// assert errors that should never occur
assert(status != U_BUFFER_OVERFLOW_ERROR);
...
}
**Looking at the thread with clrstack shows an inconsistent stack as well. We have managed code on the stack with no native code below.**
libc.so.6!__GI_raise + 57 at /build/buildd/eglibc-2.19/signal/../nptl/sysdeps/unix/sysv/linux/raise.c:56
libc.so.6!__GI_abort + 328 at /build/buildd/eglibc-2.19/stdlib/abort.c:91
libc.so.6!__assert_fail_base + 294 at /build/buildd/eglibc-2.19/assert/assert.c:92
libc.so.6 + -1
System.Globalization.Native.so!UErrorCodeToBool(UErrorCode) + 88 at /root/coreclr/src/corefx/System.Globalization.Native/locale.cpp:21
System.Globalization.Native.so!GlobalizationNative_GetLocaleName + 210 at /root/coreclr/src/corefx/System.Globalization.Native/locale.cpp:130
[InlinedCallFrame: 00007f68d8d68bb0] mscorlib.ni.dll!Interop+GlobalizationInterop.GetLocaleName(System.String, System.Text.StringBuilder, Int32)
[InlinedCallFrame: 00007f68d8d68bb0] mscorlib.ni.dll!Interop+GlobalizationInterop.GetLocaleName(System.String, System.Text.StringBuilder, Int32)
mscorlib.ni.dll!DomainNeutralILStubClass.IL_STUB_PInvoke(System.String, System.Text.StringBuilder, Int32) + 389
mscorlib.ni.dll!System.Globalization.CultureData.GetLocaleName(System.String, System.String ByRef) + 53
mscorlib.ni.dll!System.Globalization.CultureData.InitCultureData() + 302
mscorlib.ni.dll!System.Globalization.CultureData.CreateCultureData(System.String, Boolean) + 120
mscorlib.ni.dll!System.Globalization.CultureData.GetCultureData(System.String, Boolean) + 356
mscorlib.ni.dll!System.Globalization.CultureInfo..ctor(System.String, Boolean) + 39
CultureInfoConstructor2.dll!System.Globalization.Tests.CultureInfoConstructor2.<TestInvalidCulture>b__2() + 82
xunit.assert.dll!Xunit.Assert.RecordException(System.Action) + 30
xunit.assert.dll!Xunit.Assert.Throws[[System.__Canon, mscorlib]](System.Action) + 59
CultureInfoConstructor2.dll!System.Globalization.Tests.CultureInfoConstructor2.TestInvalidCulture() + 161
projectk-23923-00-xlsh8-amd64chk_01A2.exe!stress.generated.UnitTests.UT2() + 74
stress.execution.dll!stress.execution.UnitTest.Execute() + 168
stress.execution.dll!stress.execution.DedicatedThreadWorkerStrategy.RunWorker(stress.execution.ITestPattern, System.Threading.CancellationToken) + 110
stress.execution.dll!stress.execution.DedicatedThreadWorkerStrategy+<>c__DisplayClass1_0.<SpawnWorker>b__0() + 113
mscorlib.ni.dll!System.Threading.Tasks.Task.Execute() + 70
mscorlib.ni.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) + 149
mscorlib.ni.dll!System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef) + 660
mscorlib.ni.dll!System.Threading.Tasks.Task.ExecuteEntry(Boolean) + 152
mscorlib.ni.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) + 149
[GCFrame: 00007f68d8d69648]
[DebuggerU2MCatchHandlerFrame: 00007f68d8d69b38]
**Looking at the thread natively through the debugger frame 7 has an invalid ip, so it seems either the stack or unwind info has become corrupt.**
(lldb) bt
* thread #1: tid = 23459, 0x00007f68e1652cc9 libc.so.6`__GI_raise(sig=6) + 57 at raise.c:56, name = 'corerun', stop reason = signal SIGABRT
frame #0: 0x00007f68e1652cc9 libc.so.6`__GI_raise(sig=6) + 57 at raise.c:56
frame #1: 0x00007f68e16560d8 libc.so.6`__GI_abort + 328 at abort.c:89
frame #2: 0x00007f68e164bb86 libc.so.6`__assert_fail_base(fmt=<unavailable>, assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) + 294 at assert.c:92
frame #3: 0x00007f68e164bc32 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) + 66 at assert.c:101
frame #4: 0x00007f68dd5d3ed8 System.Globalization.Native.so`UErrorCodeToBool(status=U_BUFFER_OVERFLOW_ERROR) + 88 at locale.cpp:20
* frame #5: 0x00007f68dd5d4302 System.Globalization.Native.so`::GlobalizationNative_GetLocaleName(localeName=0x00007f68480300f4, value=0x00007f68d8d68a60, valueLength=157) + 210 at locale.cpp:130
frame #6: 0x00007f6866fefa45
frame #7: 0x9abcdef012345678 | True | SIGABRT_System.Globalization.Native.so!UErrorCodeToBool - **The notes in this bug are referencing the dump [projectk-23923-00-xlsh8-amd64chk_01A2.zip](https://dotnetjobresults.blob.core.windows.net/projectk-23923-00-xlsh8-stress/f7b98499-069d-4c64-b420-af5c211233a3/stress/generated/00_01hr/projectk-23923-00-xlsh8-amd64chk_01A2/projectk-23923-00-xlsh8-amd64chk_01A2.csproj/projectk-23923-00-xlsh8-amd64chk_01A2.zip?sv=2015-04-05&sr=c&sig=O%2FEO8ALizJpZh%2Bmm6lQNB%2BPinsMWLz17JO24BdFDg4U%3D&se=2017-03-29T07:34:00Z&sp=rw), more dumps are availabie for this issue if needed**
**This failure is due to an assert in System.Globalization.Native, because a call to GetLocal returned a status of U_BUFFER_OVERFLOW_ERROR. However there seems to be more going on in this dump, as the stack of this thread seems to be torn either with a corrupted stack or corrupted unwind info.**
STOP_REASON:
SIGABRT
FAILURE_HASH:
SIGABRT_System.Globalization.Native.so!UErrorCodeToBool
FAULT_STACK:
libc.so.6!__GI_raise
libc.so.6!__GI_abort
libc.so.6!__assert_fail_base
libc.so.6!__GI___assert_fail
System.Globalization.Native.so!UErrorCodeToBool
System.Globalization.Native.so!GlobalizationNative_GetLocaleName
mscorlib.ni!DomainNeutralILStubClass.IL_STUB_PInvoke
UNKNOWN!UNKNOWN
FAULT_SYMBOL:
System.Globalization.Native.so!UErrorCodeToBool
**From looking at the values from frame 5 and the code we can see that we are asserting because a status of U_BUFFER_OVERFLOW_ERROR was returned from GetLocal**
(lldb) fr s 5
frame #5: 0x00007f68dd5d4302 System.Globalization.Native.so`::GlobalizationNative_GetLocaleName(localeName=0x00007f68480300f4, value=0x00007f68d8d68a60, valueLength=157) + 210 at locale.cpp:130
(lldb) fr v
(const UChar *) localeName = 0x00007f68480300f4
(int32_t) valueLength = 157
(UChar *) value = 0x00007f68d8d68a60
(UErrorCode) status = U_BUFFER_OVERFLOW_ERROR
(char [157]) localeNameBuffer = "notavalidculture"
extern "C" int32_t GlobalizationNative_GetLocaleName(const UChar* localeName, UChar* value, int32_t valueLength)
{
UErrorCode status = U_ZERO_ERROR;
char localeNameBuffer[ULOC_FULLNAME_CAPACITY];
GetLocale(localeName, localeNameBuffer, ULOC_FULLNAME_CAPACITY, true, &status);
...
return UErrorCodeToBool(status);
}
int32_t UErrorCodeToBool(UErrorCode status)
{
...
// assert errors that should never occur
assert(status != U_BUFFER_OVERFLOW_ERROR);
...
}
**Looking at the thread with clrstack shows an inconsistent stack as well. We have managed code on the stack with no native code below.**
libc.so.6!__GI_raise + 57 at /build/buildd/eglibc-2.19/signal/../nptl/sysdeps/unix/sysv/linux/raise.c:56
libc.so.6!__GI_abort + 328 at /build/buildd/eglibc-2.19/stdlib/abort.c:91
libc.so.6!__assert_fail_base + 294 at /build/buildd/eglibc-2.19/assert/assert.c:92
libc.so.6 + -1
System.Globalization.Native.so!UErrorCodeToBool(UErrorCode) + 88 at /root/coreclr/src/corefx/System.Globalization.Native/locale.cpp:21
System.Globalization.Native.so!GlobalizationNative_GetLocaleName + 210 at /root/coreclr/src/corefx/System.Globalization.Native/locale.cpp:130
[InlinedCallFrame: 00007f68d8d68bb0] mscorlib.ni.dll!Interop+GlobalizationInterop.GetLocaleName(System.String, System.Text.StringBuilder, Int32)
[InlinedCallFrame: 00007f68d8d68bb0] mscorlib.ni.dll!Interop+GlobalizationInterop.GetLocaleName(System.String, System.Text.StringBuilder, Int32)
mscorlib.ni.dll!DomainNeutralILStubClass.IL_STUB_PInvoke(System.String, System.Text.StringBuilder, Int32) + 389
mscorlib.ni.dll!System.Globalization.CultureData.GetLocaleName(System.String, System.String ByRef) + 53
mscorlib.ni.dll!System.Globalization.CultureData.InitCultureData() + 302
mscorlib.ni.dll!System.Globalization.CultureData.CreateCultureData(System.String, Boolean) + 120
mscorlib.ni.dll!System.Globalization.CultureData.GetCultureData(System.String, Boolean) + 356
mscorlib.ni.dll!System.Globalization.CultureInfo..ctor(System.String, Boolean) + 39
CultureInfoConstructor2.dll!System.Globalization.Tests.CultureInfoConstructor2.<TestInvalidCulture>b__2() + 82
xunit.assert.dll!Xunit.Assert.RecordException(System.Action) + 30
xunit.assert.dll!Xunit.Assert.Throws[[System.__Canon, mscorlib]](System.Action) + 59
CultureInfoConstructor2.dll!System.Globalization.Tests.CultureInfoConstructor2.TestInvalidCulture() + 161
projectk-23923-00-xlsh8-amd64chk_01A2.exe!stress.generated.UnitTests.UT2() + 74
stress.execution.dll!stress.execution.UnitTest.Execute() + 168
stress.execution.dll!stress.execution.DedicatedThreadWorkerStrategy.RunWorker(stress.execution.ITestPattern, System.Threading.CancellationToken) + 110
stress.execution.dll!stress.execution.DedicatedThreadWorkerStrategy+<>c__DisplayClass1_0.<SpawnWorker>b__0() + 113
mscorlib.ni.dll!System.Threading.Tasks.Task.Execute() + 70
mscorlib.ni.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) + 149
mscorlib.ni.dll!System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef) + 660
mscorlib.ni.dll!System.Threading.Tasks.Task.ExecuteEntry(Boolean) + 152
mscorlib.ni.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) + 149
[GCFrame: 00007f68d8d69648]
[DebuggerU2MCatchHandlerFrame: 00007f68d8d69b38]
**Looking at the thread natively through the debugger frame 7 has an invalid ip, so it seems either the stack or unwind info has become corrupt.**
(lldb) bt
* thread #1: tid = 23459, 0x00007f68e1652cc9 libc.so.6`__GI_raise(sig=6) + 57 at raise.c:56, name = 'corerun', stop reason = signal SIGABRT
frame #0: 0x00007f68e1652cc9 libc.so.6`__GI_raise(sig=6) + 57 at raise.c:56
frame #1: 0x00007f68e16560d8 libc.so.6`__GI_abort + 328 at abort.c:89
frame #2: 0x00007f68e164bb86 libc.so.6`__assert_fail_base(fmt=<unavailable>, assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) + 294 at assert.c:92
frame #3: 0x00007f68e164bc32 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) + 66 at assert.c:101
frame #4: 0x00007f68dd5d3ed8 System.Globalization.Native.so`UErrorCodeToBool(status=U_BUFFER_OVERFLOW_ERROR) + 88 at locale.cpp:20
* frame #5: 0x00007f68dd5d4302 System.Globalization.Native.so`::GlobalizationNative_GetLocaleName(localeName=0x00007f68480300f4, value=0x00007f68d8d68a60, valueLength=157) + 210 at locale.cpp:130
frame #6: 0x00007f6866fefa45
frame #7: 0x9abcdef012345678 | non_priority | sigabrt system globalization native so uerrorcodetobool the notes in this bug are referencing the dump more dumps are availabie for this issue if needed this failure is due to an assert in system globalization native because a call to getlocal returned a status of u buffer overflow error however there seems to be more going on in this dump as the stack of this thread seems to be torn either with a corrupted stack or corrupted unwind info stop reason sigabrt failure hash sigabrt system globalization native so uerrorcodetobool fault stack libc so gi raise libc so gi abort libc so assert fail base libc so gi assert fail system globalization native so uerrorcodetobool system globalization native so globalizationnative getlocalename mscorlib ni domainneutralilstubclass il stub pinvoke unknown unknown fault symbol system globalization native so uerrorcodetobool from looking at the values from frame and the code we can see that we are asserting because a status of u buffer overflow error was returned from getlocal lldb fr s frame system globalization native so globalizationnative getlocalename localename value valuelength at locale cpp lldb fr v const uchar localename t valuelength uchar value uerrorcode status u buffer overflow error char localenamebuffer notavalidculture extern c t globalizationnative getlocalename const uchar localename uchar value t valuelength uerrorcode status u zero error char localenamebuffer getlocale localename localenamebuffer uloc fullname capacity true status return uerrorcodetobool status t uerrorcodetobool uerrorcode status assert errors that should never occur assert status u buffer overflow error looking at the thread with clrstack shows an inconsistent stack as well we have managed code on the stack with no native code below libc so gi raise at build buildd eglibc signal nptl sysdeps unix sysv linux raise c libc so gi abort at build buildd eglibc stdlib abort c libc so assert fail base at build buildd eglibc assert assert c libc so system globalization native so uerrorcodetobool uerrorcode at root coreclr src corefx system globalization native locale cpp system globalization native so globalizationnative getlocalename at root coreclr src corefx system globalization native locale cpp mscorlib ni dll interop globalizationinterop getlocalename system string system text stringbuilder mscorlib ni dll interop globalizationinterop getlocalename system string system text stringbuilder mscorlib ni dll domainneutralilstubclass il stub pinvoke system string system text stringbuilder mscorlib ni dll system globalization culturedata getlocalename system string system string byref mscorlib ni dll system globalization culturedata initculturedata mscorlib ni dll system globalization culturedata createculturedata system string boolean mscorlib ni dll system globalization culturedata getculturedata system string boolean mscorlib ni dll system globalization cultureinfo ctor system string boolean dll system globalization tests b xunit assert dll xunit assert recordexception system action xunit assert dll xunit assert throws system action dll system globalization tests testinvalidculture projectk exe stress generated unittests stress execution dll stress execution unittest execute stress execution dll stress execution dedicatedthreadworkerstrategy runworker stress execution itestpattern system threading cancellationtoken stress execution dll stress execution dedicatedthreadworkerstrategy c b mscorlib ni dll system threading tasks task execute mscorlib ni dll system threading executioncontext run system threading executioncontext system threading contextcallback system object mscorlib ni dll system threading tasks task executewiththreadlocal system threading tasks task byref mscorlib ni dll system threading tasks task executeentry boolean mscorlib ni dll system threading executioncontext run system threading executioncontext system threading contextcallback system object looking at the thread natively through the debugger frame has an invalid ip so it seems either the stack or unwind info has become corrupt lldb bt thread tid libc so gi raise sig at raise c name corerun stop reason signal sigabrt frame libc so gi raise sig at raise c frame libc so gi abort at abort c frame libc so assert fail base fmt assertion file line function at assert c frame libc so gi assert fail assertion file line function at assert c frame system globalization native so uerrorcodetobool status u buffer overflow error at locale cpp frame system globalization native so globalizationnative getlocalename localename value valuelength at locale cpp frame frame | 0 |
36,988 | 12,463,925,945 | IssuesEvent | 2020-05-28 11:31:56 | corona-warn-app/cwa-server | https://api.github.com/repos/corona-warn-app/cwa-server | closed | Add correct Verification Key ID and Version | community enhancement further input needed security | ## Current Implementation
The Verification Key ID and Verification Key Version inside the "Signature Info" are blank. See [here](https://github.com/corona-warn-app/cwa-server/pull/240/files#diff-a5805e5e55bd745458a2dfac8cf91938R88).
## Suggested Enhancement
Add the correct Verification Key ID and Verification Key Version.
## Details
These values will eventually be provided by Google/Apple, when we send them our prod certificate. This issue is blocked until that happens. | True | Add correct Verification Key ID and Version - ## Current Implementation
The Verification Key ID and Verification Key Version inside the "Signature Info" are blank. See [here](https://github.com/corona-warn-app/cwa-server/pull/240/files#diff-a5805e5e55bd745458a2dfac8cf91938R88).
## Suggested Enhancement
Add the correct Verification Key ID and Verification Key Version.
## Details
These values will eventually be provided by Google/Apple, when we send them our prod certificate. This issue is blocked until that happens. | non_priority | add correct verification key id and version current implementation the verification key id and verification key version inside the signature info are blank see suggested enhancement add the correct verification key id and verification key version details these values will eventually be provided by google apple when we send them our prod certificate this issue is blocked until that happens | 0 |
137,934 | 30,778,211,193 | IssuesEvent | 2023-07-31 08:11:35 | AntaresSimulatorTeam/Antares_Simulator | https://api.github.com/repos/AntaresSimulatorTeam/Antares_Simulator | closed | Memory allocation, use std::vector | cleaning code quality | MemAlloc is still used in those files, replace it with std::vector
- [ ] simulation/sim_allocation_tableaux.cpp
- [ ] optimisation/opt_alloc_probleme_a_optimiser.cpp
- [ ] optimisation/opt_nombre_min_groupes_demarres_couts_demarrage.cpp | 1.0 | Memory allocation, use std::vector - MemAlloc is still used in those files, replace it with std::vector
- [ ] simulation/sim_allocation_tableaux.cpp
- [ ] optimisation/opt_alloc_probleme_a_optimiser.cpp
- [ ] optimisation/opt_nombre_min_groupes_demarres_couts_demarrage.cpp | non_priority | memory allocation use std vector memalloc is still used in those files replace it with std vector simulation sim allocation tableaux cpp optimisation opt alloc probleme a optimiser cpp optimisation opt nombre min groupes demarres couts demarrage cpp | 0 |
426,300 | 29,515,183,378 | IssuesEvent | 2023-06-04 12:10:24 | qwer7824/PharmacyRecommendation | https://api.github.com/repos/qwer7824/PharmacyRecommendation | closed | Spock 프레임워크를 이용한 테스트 작성 | documentation enhancement | Spock 란 Groovy 언어를 이용하여 테스트 코드를 작성할 수 있는 프레임워크이며 JUnit 과 비교하여 코드를 더 간결하게 작성가능하다.
테스트 메소드 이름을 문자열로 작성할 수 있으며 Given , When , Then 코드 블럭을 명확히 구분한다.
Junit 의 경우 주석으로 블록을 구분했었고 메소드 이름 또한 제약사항이 존재한다.
---
테스트 클래스는 Groovy 클래스로 생성하고 , Specification 클래스를 상속 받는다.
feature(테스트 메서드는) def를 이용하여 함수로 선언하며, 하나 이상 블록이 존재해야함
given 블록 : 테스트에 필요한 값을 준비한다.
when 블록 : 테스트할 코드를 실행한다.
then 블록 : when과 함께 사용하며 예외 및 결과 값을 검증한다.
expect 블록 : then 과 같으며 when 을 필요로 하지 않기 때문에 간단한 테스트 및 where와 같이 사용된다.
where 블록 : 데이터가 다르고 로직이 동일한 경우 동일한 테스트에 대한 중복 코드 제거가능하다. | 1.0 | Spock 프레임워크를 이용한 테스트 작성 - Spock 란 Groovy 언어를 이용하여 테스트 코드를 작성할 수 있는 프레임워크이며 JUnit 과 비교하여 코드를 더 간결하게 작성가능하다.
테스트 메소드 이름을 문자열로 작성할 수 있으며 Given , When , Then 코드 블럭을 명확히 구분한다.
Junit 의 경우 주석으로 블록을 구분했었고 메소드 이름 또한 제약사항이 존재한다.
---
테스트 클래스는 Groovy 클래스로 생성하고 , Specification 클래스를 상속 받는다.
feature(테스트 메서드는) def를 이용하여 함수로 선언하며, 하나 이상 블록이 존재해야함
given 블록 : 테스트에 필요한 값을 준비한다.
when 블록 : 테스트할 코드를 실행한다.
then 블록 : when과 함께 사용하며 예외 및 결과 값을 검증한다.
expect 블록 : then 과 같으며 when 을 필요로 하지 않기 때문에 간단한 테스트 및 where와 같이 사용된다.
where 블록 : 데이터가 다르고 로직이 동일한 경우 동일한 테스트에 대한 중복 코드 제거가능하다. | non_priority | spock 프레임워크를 이용한 테스트 작성 spock 란 groovy 언어를 이용하여 테스트 코드를 작성할 수 있는 프레임워크이며 junit 과 비교하여 코드를 더 간결하게 작성가능하다 테스트 메소드 이름을 문자열로 작성할 수 있으며 given when then 코드 블럭을 명확히 구분한다 junit 의 경우 주석으로 블록을 구분했었고 메소드 이름 또한 제약사항이 존재한다 테스트 클래스는 groovy 클래스로 생성하고 specification 클래스를 상속 받는다 feature 테스트 메서드는 def를 이용하여 함수로 선언하며 하나 이상 블록이 존재해야함 given 블록 테스트에 필요한 값을 준비한다 when 블록 테스트할 코드를 실행한다 then 블록 when과 함께 사용하며 예외 및 결과 값을 검증한다 expect 블록 then 과 같으며 when 을 필요로 하지 않기 때문에 간단한 테스트 및 where와 같이 사용된다 where 블록 데이터가 다르고 로직이 동일한 경우 동일한 테스트에 대한 중복 코드 제거가능하다 | 0 |
338,482 | 24,587,569,678 | IssuesEvent | 2022-10-13 21:16:27 | chakra-ui/chakra-ui-docs | https://api.github.com/repos/chakra-ui/chakra-ui-docs | closed | Theming: Avatar | good first issue created-by: Chakra UI team Topic: Documentation 📚 hacktoberfest | ### Subject
Avatar
### Description
Create a basic theming documentation for the `Avatar` component that showcases how you can style it.
Theming file: [Avatar theming file](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/avatar/theming.mdx)
### Template Reference
For single part components, feel free to use the `Button` component theming doc as a template reference: [Button theming](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/button/theming.mdx).
For multipart components, feel free to use the `Input` component theming doc as a template reference: [Input theming](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/input/theming.mdx). | 1.0 | Theming: Avatar - ### Subject
Avatar
### Description
Create a basic theming documentation for the `Avatar` component that showcases how you can style it.
Theming file: [Avatar theming file](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/avatar/theming.mdx)
### Template Reference
For single part components, feel free to use the `Button` component theming doc as a template reference: [Button theming](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/button/theming.mdx).
For multipart components, feel free to use the `Input` component theming doc as a template reference: [Input theming](https://github.com/chakra-ui/chakra-ui-docs/blob/main/content/docs/components/input/theming.mdx). | non_priority | theming avatar subject avatar description create a basic theming documentation for the avatar component that showcases how you can style it theming file template reference for single part components feel free to use the button component theming doc as a template reference for multipart components feel free to use the input component theming doc as a template reference | 0 |
18,230 | 4,240,007,956 | IssuesEvent | 2016-07-06 11:50:04 | openucx/ucx | https://api.github.com/repos/openucx/ucx | opened | UCX testing procedure | documentation | Please add a short text file to ucx/doc, or to ucx wiki, to describe the mandatory procedure for testing.
For example:
- Every change should have a unit test
- Which tests are running now (Jenkins, gtest, cov, valgrind,...)
- When to test on IMB large scale
o E,g run a benchmark (IMB for example) on one host with ppn=16 with UCX_SHM_DEVICES=posix
- MTT
| 1.0 | UCX testing procedure - Please add a short text file to ucx/doc, or to ucx wiki, to describe the mandatory procedure for testing.
For example:
- Every change should have a unit test
- Which tests are running now (Jenkins, gtest, cov, valgrind,...)
- When to test on IMB large scale
o E,g run a benchmark (IMB for example) on one host with ppn=16 with UCX_SHM_DEVICES=posix
- MTT
| non_priority | ucx testing procedure please add a short text file to ucx doc or to ucx wiki to describe the mandatory procedure for testing for example every change should have a unit test which tests are running now jenkins gtest cov valgrind when to test on imb large scale o e g run a benchmark imb for example on one host with ppn with ucx shm devices posix mtt | 0 |
54,948 | 13,943,430,981 | IssuesEvent | 2020-10-22 23:06:11 | GooseWSS/node-express-realworld-example-app | https://api.github.com/repos/GooseWSS/node-express-realworld-example-app | opened | WS-2019-0257 (Medium) detected in csv-parse-1.3.3.tgz | security vulnerability | ## WS-2019-0257 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>csv-parse-1.3.3.tgz</b></p></summary>
<p>CSV parsing implementing the Node.js `stream.Transform` API</p>
<p>Library home page: <a href="https://registry.npmjs.org/csv-parse/-/csv-parse-1.3.3.tgz">https://registry.npmjs.org/csv-parse/-/csv-parse-1.3.3.tgz</a></p>
<p>Path to dependency file: node-express-realworld-example-app/package.json</p>
<p>Path to vulnerable library: node-express-realworld-example-app/node_modules/csv-parse/package.json</p>
<p>
Dependency Hierarchy:
- newman-3.10.0.tgz (Root Library)
- :x: **csv-parse-1.3.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GooseWSS/node-express-realworld-example-app/commit/532d31bcb4a1ee884633f7e684aab7c5bb204ca5">532d31bcb4a1ee884633f7e684aab7c5bb204ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Csv-parse prior to 4.4.6 is vulnerable to Regular Expression Denial of Service. The __isInt() function contains a malformed regular expression that processes large specially-crafted input very slowly, this leads to a Denial of Service which is triggered when using the cast option.
<p>Publish Date: 2019-09-17
<p>URL: <a href=https://github.com/adaltas/node-csv-parse/commit/b9d35940c6815cdf1dfd6b21857a1f6d0fd51e4a>WS-2019-0257</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1171">https://www.npmjs.com/advisories/1171</a></p>
<p>Release Date: 2019-09-24</p>
<p>Fix Resolution: 4.4.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"csv-parse","packageVersion":"1.3.3","isTransitiveDependency":true,"dependencyTree":"newman:3.10.0;csv-parse:1.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.4.6"}],"vulnerabilityIdentifier":"WS-2019-0257","vulnerabilityDetails":"Csv-parse prior to 4.4.6 is vulnerable to Regular Expression Denial of Service. The __isInt() function contains a malformed regular expression that processes large specially-crafted input very slowly, this leads to a Denial of Service which is triggered when using the cast option.","vulnerabilityUrl":"https://github.com/adaltas/node-csv-parse/commit/b9d35940c6815cdf1dfd6b21857a1f6d0fd51e4a","cvss2Severity":"medium","cvss2Score":"6.5","extraData":{}}</REMEDIATE> --> | True | WS-2019-0257 (Medium) detected in csv-parse-1.3.3.tgz - ## WS-2019-0257 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>csv-parse-1.3.3.tgz</b></p></summary>
<p>CSV parsing implementing the Node.js `stream.Transform` API</p>
<p>Library home page: <a href="https://registry.npmjs.org/csv-parse/-/csv-parse-1.3.3.tgz">https://registry.npmjs.org/csv-parse/-/csv-parse-1.3.3.tgz</a></p>
<p>Path to dependency file: node-express-realworld-example-app/package.json</p>
<p>Path to vulnerable library: node-express-realworld-example-app/node_modules/csv-parse/package.json</p>
<p>
Dependency Hierarchy:
- newman-3.10.0.tgz (Root Library)
- :x: **csv-parse-1.3.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GooseWSS/node-express-realworld-example-app/commit/532d31bcb4a1ee884633f7e684aab7c5bb204ca5">532d31bcb4a1ee884633f7e684aab7c5bb204ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Csv-parse prior to 4.4.6 is vulnerable to Regular Expression Denial of Service. The __isInt() function contains a malformed regular expression that processes large specially-crafted input very slowly, this leads to a Denial of Service which is triggered when using the cast option.
<p>Publish Date: 2019-09-17
<p>URL: <a href=https://github.com/adaltas/node-csv-parse/commit/b9d35940c6815cdf1dfd6b21857a1f6d0fd51e4a>WS-2019-0257</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1171">https://www.npmjs.com/advisories/1171</a></p>
<p>Release Date: 2019-09-24</p>
<p>Fix Resolution: 4.4.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"csv-parse","packageVersion":"1.3.3","isTransitiveDependency":true,"dependencyTree":"newman:3.10.0;csv-parse:1.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.4.6"}],"vulnerabilityIdentifier":"WS-2019-0257","vulnerabilityDetails":"Csv-parse prior to 4.4.6 is vulnerable to Regular Expression Denial of Service. The __isInt() function contains a malformed regular expression that processes large specially-crafted input very slowly, this leads to a Denial of Service which is triggered when using the cast option.","vulnerabilityUrl":"https://github.com/adaltas/node-csv-parse/commit/b9d35940c6815cdf1dfd6b21857a1f6d0fd51e4a","cvss2Severity":"medium","cvss2Score":"6.5","extraData":{}}</REMEDIATE> --> | non_priority | ws medium detected in csv parse tgz ws medium severity vulnerability vulnerable library csv parse tgz csv parsing implementing the node js stream transform api library home page a href path to dependency file node express realworld example app package json path to vulnerable library node express realworld example app node modules csv parse package json dependency hierarchy newman tgz root library x csv parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details csv parse prior to is vulnerable to regular expression denial of service the isint function contains a malformed regular expression that processes large specially crafted input very slowly this leads to a denial of service which is triggered when using the cast option publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails csv parse prior to is vulnerable to regular expression denial of service the isint function contains a malformed regular expression that processes large specially crafted input very slowly this leads to a denial of service which is triggered when using the cast option vulnerabilityurl | 0 |
32,538 | 7,545,061,550 | IssuesEvent | 2018-04-17 20:25:06 | Motoxpro/WorldCupStatsSite | https://api.github.com/repos/Motoxpro/WorldCupStatsSite | opened | Design/Implement results table JSON | Code/Backend | Create the json that contains the number of splits, and the right format ( (2) 2:12.312 +12.000 ) for the table | 1.0 | Design/Implement results table JSON - Create the json that contains the number of splits, and the right format ( (2) 2:12.312 +12.000 ) for the table | non_priority | design implement results table json create the json that contains the number of splits and the right format for the table | 0 |
2,308 | 2,909,472,825 | IssuesEvent | 2015-06-20 19:29:41 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | New error metadata system required. | A-build I-enhancement | It'd be nice to hang on to the [error index][err-idx] in the Rust 1.2 release... currently it's disabled due to the flakiness of the metadata collation system (see #25706).
Amongst the issues with the previous implementation:
* Unsynchronised concurrent access of files, leading to the use of malformed JSON data.
* No way to determine the freshness of metadata files.
* Failure to write files (not thoroughly investigated). Probably related to stale files hanging around.
Before I go ahead and implement a crazy new system, I just want to check the sanity of my design:
* Alter the `__build_diagnostic_array!` invocations so that the identifier used for diagnostics arrays is something unlikely to be used elsewhere. This step isn't strictly necessary, as the current name `DIAGNOSTICS` is already quite unique.
* Register an internal lint (i.e. one that only runs against the compiler's source) that extracts error metadata from every item with the unique name used by `__build_diagnostic_array!`. Here we run into the same potential concurrency problem as before, because crates can be compiled in parallel. As long as the lint doesn't attempt to check uniqueness by reading files for *other* crates, the parallel writes to separate per-crate files should be ok. E.g. when running the lint against `librustc_typeck`, dump metadata to `tmp/extended-errors/librustc_typeck.json`. What worries me about this step is that it is essentially the same as dumping metadata from the `__build_diagnostic_array!` without uniqueness-checks, which we tried **unsuccessfully** (that's the unknown error).
* In the error-index-generator - which runs once all crates have been compiled - we can check uniqueness.
I still feel like this is very similar to the previous, failing system, and would like to alter it more drastically if anyone has any ideas...
cc @pnkfelix @alexcrichton @Manishearth
[err-idx]: http://doc.rust-lang.org/error-index.html | 1.0 | New error metadata system required. - It'd be nice to hang on to the [error index][err-idx] in the Rust 1.2 release... currently it's disabled due to the flakiness of the metadata collation system (see #25706).
Amongst the issues with the previous implementation:
* Unsynchronised concurrent access of files, leading to the use of malformed JSON data.
* No way to determine the freshness of metadata files.
* Failure to write files (not thoroughly investigated). Probably related to stale files hanging around.
Before I go ahead and implement a crazy new system, I just want to check the sanity of my design:
* Alter the `__build_diagnostic_array!` invocations so that the identifier used for diagnostics arrays is something unlikely to be used elsewhere. This step isn't strictly necessary, as the current name `DIAGNOSTICS` is already quite unique.
* Register an internal lint (i.e. one that only runs against the compiler's source) that extracts error metadata from every item with the unique name used by `__build_diagnostic_array!`. Here we run into the same potential concurrency problem as before, because crates can be compiled in parallel. As long as the lint doesn't attempt to check uniqueness by reading files for *other* crates, the parallel writes to separate per-crate files should be ok. E.g. when running the lint against `librustc_typeck`, dump metadata to `tmp/extended-errors/librustc_typeck.json`. What worries me about this step is that it is essentially the same as dumping metadata from the `__build_diagnostic_array!` without uniqueness-checks, which we tried **unsuccessfully** (that's the unknown error).
* In the error-index-generator - which runs once all crates have been compiled - we can check uniqueness.
I still feel like this is very similar to the previous, failing system, and would like to alter it more drastically if anyone has any ideas...
cc @pnkfelix @alexcrichton @Manishearth
[err-idx]: http://doc.rust-lang.org/error-index.html | non_priority | new error metadata system required it d be nice to hang on to the in the rust release currently it s disabled due to the flakiness of the metadata collation system see amongst the issues with the previous implementation unsynchronised concurrent access of files leading to the use of malformed json data no way to determine the freshness of metadata files failure to write files not thoroughly investigated probably related to stale files hanging around before i go ahead and implement a crazy new system i just want to check the sanity of my design alter the build diagnostic array invocations so that the identifier used for diagnostics arrays is something unlikely to be used elsewhere this step isn t strictly necessary as the current name diagnostics is already quite unique register an internal lint i e one that only runs against the compiler s source that extracts error metadata from every item with the unique name used by build diagnostic array here we run into the same potential concurrency problem as before because crates can be compiled in parallel as long as the lint doesn t attempt to check uniqueness by reading files for other crates the parallel writes to separate per crate files should be ok e g when running the lint against librustc typeck dump metadata to tmp extended errors librustc typeck json what worries me about this step is that it is essentially the same as dumping metadata from the build diagnostic array without uniqueness checks which we tried unsuccessfully that s the unknown error in the error index generator which runs once all crates have been compiled we can check uniqueness i still feel like this is very similar to the previous failing system and would like to alter it more drastically if anyone has any ideas cc pnkfelix alexcrichton manishearth | 0 |
113,336 | 14,410,349,204 | IssuesEvent | 2020-12-04 04:37:44 | cammelworks/doubleEdged | https://api.github.com/repos/cammelworks/doubleEdged | opened | 新規モデルについて | Design idea | #概要
## キャラクターのバリエーションを増やす
- DoubleEdgedの世界観に合うようなキャラを増やす
- 白/黒のペアでセット
- 例)炭鉱の隊長さん
## 新規フィールド
- 洞窟
- 崖付近などなど | 1.0 | 新規モデルについて - #概要
## キャラクターのバリエーションを増やす
- DoubleEdgedの世界観に合うようなキャラを増やす
- 白/黒のペアでセット
- 例)炭鉱の隊長さん
## 新規フィールド
- 洞窟
- 崖付近などなど | non_priority | 新規モデルについて 概要 キャラクターのバリエーションを増やす doubleedgedの世界観に合うようなキャラを増やす 白 黒のペアでセット 例 炭鉱の隊長さん 新規フィールド 洞窟 崖付近などなど | 0 |
46,835 | 24,748,509,570 | IssuesEvent | 2022-10-21 11:50:30 | layer5io/layer5 | https://api.github.com/repos/layer5io/layer5 | opened | [CI] Performance: Configure Lighthouse-CI GitHub Action | kind/enhancement help wanted area/ci kind/performance | #### Current Behavior
The layer5.io site has performance challenges.
#### Desired Behavior
Ideally, the https://github.com/GoogleChrome/lighthouse-ci GitHub Action runs on open of each PR.
---
<img src="https://raw.githubusercontent.com/layer5io/layer5/master/.github/assets/images/layer5/5-light-small.svg" width="16px" align="left" /><h3> Contributor Resources and <a href="https://layer5.io/community/handbook">Handbook</a></h3>
The layer5.io website uses Gatsby, React, and GitHub Pages. Site content is found under the [`master` branch](https://github.com/layer5io/layer5/tree/master).
- 📚 See [contributing instructions](https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md)
- 🎨 Wireframes and designs for Layer5 site in [Figma](https://www.figma.com/file/5ZwEkSJwUPitURD59YHMEN/Layer5-Designs).
- 🙋🏾🙋🏼 Questions: [Discussion Forum](https://discuss.layer5.io) and [Community Slack](http://slack.layer5.io)
| True | [CI] Performance: Configure Lighthouse-CI GitHub Action - #### Current Behavior
The layer5.io site has performance challenges.
#### Desired Behavior
Ideally, the https://github.com/GoogleChrome/lighthouse-ci GitHub Action runs on open of each PR.
---
<img src="https://raw.githubusercontent.com/layer5io/layer5/master/.github/assets/images/layer5/5-light-small.svg" width="16px" align="left" /><h3> Contributor Resources and <a href="https://layer5.io/community/handbook">Handbook</a></h3>
The layer5.io website uses Gatsby, React, and GitHub Pages. Site content is found under the [`master` branch](https://github.com/layer5io/layer5/tree/master).
- 📚 See [contributing instructions](https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md)
- 🎨 Wireframes and designs for Layer5 site in [Figma](https://www.figma.com/file/5ZwEkSJwUPitURD59YHMEN/Layer5-Designs).
- 🙋🏾🙋🏼 Questions: [Discussion Forum](https://discuss.layer5.io) and [Community Slack](http://slack.layer5.io)
| non_priority | performance configure lighthouse ci github action current behavior the io site has performance challenges desired behavior ideally the github action runs on open of each pr contributor resources and a href the io website uses gatsby react and github pages site content is found under the 📚 see 🎨 wireframes and designs for site in 🙋🏾🙋🏼 questions and | 0 |
65,251 | 19,297,921,356 | IssuesEvent | 2021-12-12 22:03:59 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | Self-image in portrait despite device landscape rotation | T-Defect | ### Steps to reproduce
1. Enable device rotation on Android;
2. Start a video call in landscape orientation, or rotate device to landscape after call is connected;
### Outcome
#### What did you expect?
I expected the in-picture display of my own camera to rotate to landscape …
#### What happened instead?
… but it stayed at portrait, thus cutting off part of the image. Here is a screenshot of the device in landscape mode, showing the portrait self-image:

### Your phone model
OnePlus 5T
### Operating system version
LineageOS 18.1
### Application version and app store
1.3.7 [40103070] from F-Droid
### Homeserver
matrix.madduck.net
### Will you send logs?
No | 1.0 | Self-image in portrait despite device landscape rotation - ### Steps to reproduce
1. Enable device rotation on Android;
2. Start a video call in landscape orientation, or rotate device to landscape after call is connected;
### Outcome
#### What did you expect?
I expected the in-picture display of my own camera to rotate to landscape …
#### What happened instead?
… but it stayed at portrait, thus cutting off part of the image. Here is a screenshot of the device in landscape mode, showing the portrait self-image:

### Your phone model
OnePlus 5T
### Operating system version
LineageOS 18.1
### Application version and app store
1.3.7 [40103070] from F-Droid
### Homeserver
matrix.madduck.net
### Will you send logs?
No | non_priority | self image in portrait despite device landscape rotation steps to reproduce enable device rotation on android start a video call in landscape orientation or rotate device to landscape after call is connected outcome what did you expect i expected the in picture display of my own camera to rotate to landscape … what happened instead … but it stayed at portrait thus cutting off part of the image here is a screenshot of the device in landscape mode showing the portrait self image your phone model oneplus operating system version lineageos application version and app store from f droid homeserver matrix madduck net will you send logs no | 0 |
96,970 | 12,194,951,756 | IssuesEvent | 2020-04-29 16:34:37 | Amsterdam/amsterdam-styled-components | https://api.github.com/repos/Amsterdam/amsterdam-styled-components | closed | Textarea component | design-system-alignment enhancement good first issue help wanted | Hi all,
We'd like to use the `Textarea` - component, as described in the designsystem, in our application.
https://designsystem.amsterdam.nl/7awj1hc9f/p/939dd9-text-area
Is that something that is feasible?
Thank you! | 1.0 | Textarea component - Hi all,
We'd like to use the `Textarea` - component, as described in the designsystem, in our application.
https://designsystem.amsterdam.nl/7awj1hc9f/p/939dd9-text-area
Is that something that is feasible?
Thank you! | non_priority | textarea component hi all we d like to use the textarea component as described in the designsystem in our application is that something that is feasible thank you | 0 |
41,404 | 5,355,311,754 | IssuesEvent | 2017-02-20 12:36:12 | QualiSystems/OpenStack-Shell | https://api.github.com/repos/QualiSystems/OpenStack-Shell | closed | No default route for App Connectivity Subnets | bug P1 Resolved Test Plan Ready | Right now when we create Networks and Subnets for App connectivity, default routes get added to the subnets when the interfaces come up. This results in completely messed up routing tables on VM and the connectivity to VM is potentially lost. The way to fix this is - when we create a Subnet (during connectivity) we need to explicitly specify that the particular subnet does not have a default g/w | 1.0 | No default route for App Connectivity Subnets - Right now when we create Networks and Subnets for App connectivity, default routes get added to the subnets when the interfaces come up. This results in completely messed up routing tables on VM and the connectivity to VM is potentially lost. The way to fix this is - when we create a Subnet (during connectivity) we need to explicitly specify that the particular subnet does not have a default g/w | non_priority | no default route for app connectivity subnets right now when we create networks and subnets for app connectivity default routes get added to the subnets when the interfaces come up this results in completely messed up routing tables on vm and the connectivity to vm is potentially lost the way to fix this is when we create a subnet during connectivity we need to explicitly specify that the particular subnet does not have a default g w | 0 |
30,992 | 6,388,574,418 | IssuesEvent | 2017-08-03 15:49:28 | pymc-devs/pymc3 | https://api.github.com/repos/pymc-devs/pymc3 | closed | adapt_diag initialization does not work for transformed variable | defects | Minimalistic code:
```python
with pm.Model():
mu = pm.Uniform('mu', 0., 1.)
sd = pm.HalfCauchy('sd', beta=10.)
obs = pm.Normal('y', mu=mu, sd=sd, observed=np.random.randn(100,1))
pm.sample(init='adapt_diag')
```
raise:
```python
Traceback (most recent call last):
File "<ipython-input-10-aea93a5e8087>", line 5, in <module>
pm.sample(init='adapt_diag')
File "/home/laoj/Documents/Github/pymc3/pymc3/sampling.py", line 247, in sample
progressbar=progressbar, **args)
File "/home/laoj/Documents/Github/pymc3/pymc3/sampling.py", line 729, in init_nuts
vals = distribution.draw_values(model.free_RVs)
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 194, in draw_values
values.append(_draw_value(param, point=point, givens=givens.values()))
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 258, in _draw_value
func = _compile_theano_function(param, variables)
File "/home/laoj/Documents/Github/pymc3/pymc3/memoize.py", line 16, in memoizer
cache[key] = obj(*args, **kwargs)
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 220, in _compile_theano_function
allow_input_downcast=True)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function.py", line 326, in function
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/pfunc.py", line 486, in pfunc
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 1808, in orig_function
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 1446, in __init__
accept_inplace)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 177, in std_fgraph
update_mapping=update_mapping)
File "/usr/local/lib/python3.5/dist-packages/theano/gof/fg.py", line 175, in __init__
self.__import_r__(output, reason="init")
File "/usr/local/lib/python3.5/dist-packages/theano/gof/fg.py", line 356, in __import_r__
raise MissingInputError("Undeclared input", variable=variable)
MissingInputError: Undeclared input
``` | 1.0 | adapt_diag initialization does not work for transformed variable - Minimalistic code:
```python
with pm.Model():
mu = pm.Uniform('mu', 0., 1.)
sd = pm.HalfCauchy('sd', beta=10.)
obs = pm.Normal('y', mu=mu, sd=sd, observed=np.random.randn(100,1))
pm.sample(init='adapt_diag')
```
raise:
```python
Traceback (most recent call last):
File "<ipython-input-10-aea93a5e8087>", line 5, in <module>
pm.sample(init='adapt_diag')
File "/home/laoj/Documents/Github/pymc3/pymc3/sampling.py", line 247, in sample
progressbar=progressbar, **args)
File "/home/laoj/Documents/Github/pymc3/pymc3/sampling.py", line 729, in init_nuts
vals = distribution.draw_values(model.free_RVs)
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 194, in draw_values
values.append(_draw_value(param, point=point, givens=givens.values()))
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 258, in _draw_value
func = _compile_theano_function(param, variables)
File "/home/laoj/Documents/Github/pymc3/pymc3/memoize.py", line 16, in memoizer
cache[key] = obj(*args, **kwargs)
File "/home/laoj/Documents/Github/pymc3/pymc3/distributions/distribution.py", line 220, in _compile_theano_function
allow_input_downcast=True)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function.py", line 326, in function
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/pfunc.py", line 486, in pfunc
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 1808, in orig_function
output_keys=output_keys)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 1446, in __init__
accept_inplace)
File "/usr/local/lib/python3.5/dist-packages/theano/compile/function_module.py", line 177, in std_fgraph
update_mapping=update_mapping)
File "/usr/local/lib/python3.5/dist-packages/theano/gof/fg.py", line 175, in __init__
self.__import_r__(output, reason="init")
File "/usr/local/lib/python3.5/dist-packages/theano/gof/fg.py", line 356, in __import_r__
raise MissingInputError("Undeclared input", variable=variable)
MissingInputError: Undeclared input
``` | non_priority | adapt diag initialization does not work for transformed variable minimalistic code python with pm model mu pm uniform mu sd pm halfcauchy sd beta obs pm normal y mu mu sd sd observed np random randn pm sample init adapt diag raise python traceback most recent call last file line in pm sample init adapt diag file home laoj documents github sampling py line in sample progressbar progressbar args file home laoj documents github sampling py line in init nuts vals distribution draw values model free rvs file home laoj documents github distributions distribution py line in draw values values append draw value param point point givens givens values file home laoj documents github distributions distribution py line in draw value func compile theano function param variables file home laoj documents github memoize py line in memoizer cache obj args kwargs file home laoj documents github distributions distribution py line in compile theano function allow input downcast true file usr local lib dist packages theano compile function py line in function output keys output keys file usr local lib dist packages theano compile pfunc py line in pfunc output keys output keys file usr local lib dist packages theano compile function module py line in orig function output keys output keys file usr local lib dist packages theano compile function module py line in init accept inplace file usr local lib dist packages theano compile function module py line in std fgraph update mapping update mapping file usr local lib dist packages theano gof fg py line in init self import r output reason init file usr local lib dist packages theano gof fg py line in import r raise missinginputerror undeclared input variable variable missinginputerror undeclared input | 0 |
54,205 | 13,900,902,774 | IssuesEvent | 2020-10-20 01:35:30 | gate5/angular | https://api.github.com/repos/gate5/angular | opened | CVE-2019-5413 (High) detected in morgan-1.6.1.tgz, morgan-1.9.0.tgz | security vulnerability | ## CVE-2019-5413 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>morgan-1.6.1.tgz</b>, <b>morgan-1.9.0.tgz</b></p></summary>
<p>
<details><summary><b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: angular/yarn.lock</p>
<p>Path to vulnerable library: angular/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-connect-5.0.0.tgz (Root Library)
- connect-2.30.2.tgz
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>morgan-1.9.0.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.9.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.9.0.tgz</a></p>
<p>Path to dependency file: angular/aio/yarn.lock</p>
<p>Path to vulnerable library: angular/aio/yarn.lock</p>
<p>
Dependency Hierarchy:
- firebase-tools-5.1.1.tgz (Root Library)
- superstatic-6.0.3.tgz
- :x: **morgan-1.9.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gate5/angular/commit/cf1f1c0344fa01406f61ff7437a72714be39b47e">cf1f1c0344fa01406f61ff7437a72714be39b47e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An attacker can use the format parameter to inject arbitrary commands in the npm package morgan < 1.9.1.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413>CVE-2019-5413</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390881">https://hackerone.com/reports/390881</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
| True | CVE-2019-5413 (High) detected in morgan-1.6.1.tgz, morgan-1.9.0.tgz - ## CVE-2019-5413 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>morgan-1.6.1.tgz</b>, <b>morgan-1.9.0.tgz</b></p></summary>
<p>
<details><summary><b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: angular/yarn.lock</p>
<p>Path to vulnerable library: angular/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-connect-5.0.0.tgz (Root Library)
- connect-2.30.2.tgz
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>morgan-1.9.0.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.9.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.9.0.tgz</a></p>
<p>Path to dependency file: angular/aio/yarn.lock</p>
<p>Path to vulnerable library: angular/aio/yarn.lock</p>
<p>
Dependency Hierarchy:
- firebase-tools-5.1.1.tgz (Root Library)
- superstatic-6.0.3.tgz
- :x: **morgan-1.9.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gate5/angular/commit/cf1f1c0344fa01406f61ff7437a72714be39b47e">cf1f1c0344fa01406f61ff7437a72714be39b47e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An attacker can use the format parameter to inject arbitrary commands in the npm package morgan < 1.9.1.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413>CVE-2019-5413</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390881">https://hackerone.com/reports/390881</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in morgan tgz morgan tgz cve high severity vulnerability vulnerable libraries morgan tgz morgan tgz morgan tgz http request logger middleware for node js library home page a href path to dependency file angular yarn lock path to vulnerable library angular yarn lock dependency hierarchy gulp connect tgz root library connect tgz x morgan tgz vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file angular aio yarn lock path to vulnerable library angular aio yarn lock dependency hierarchy firebase tools tgz root library superstatic tgz x morgan tgz vulnerable library found in head commit a href vulnerability details an attacker can use the format parameter to inject arbitrary commands in the npm package morgan publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
100,868 | 21,553,118,452 | IssuesEvent | 2022-04-30 01:30:46 | martinGM05/Code-Challenge | https://api.github.com/repos/martinGM05/Code-Challenge | closed | Write scripts to automate commands | Code | - [x] test - `node ./node_modules/jest/bin/jest.js`
- [x] linter - `node ./node_modules/eslint/bin/eslint.js .`
- [x] linter-fix - `node ./node_modules/eslint/bin/eslint.js . --fix`
- [x] server - `node ./lib/server.js` | 1.0 | Write scripts to automate commands - - [x] test - `node ./node_modules/jest/bin/jest.js`
- [x] linter - `node ./node_modules/eslint/bin/eslint.js .`
- [x] linter-fix - `node ./node_modules/eslint/bin/eslint.js . --fix`
- [x] server - `node ./lib/server.js` | non_priority | write scripts to automate commands test node node modules jest bin jest js linter node node modules eslint bin eslint js linter fix node node modules eslint bin eslint js fix server node lib server js | 0 |
80,742 | 10,054,211,068 | IssuesEvent | 2019-07-21 23:38:01 | ruizhengyun/designPattern-note | https://api.github.com/repos/ruizhengyun/designPattern-note | opened | Javascript 设计模式之面向对象与 UML 类图 | designPattern | # 为什么要写 Javascript 设计模式小书
关于 Javascript 设计模式的文章与书有很多很多,我写这小书主要记录我的学习过程中的笔记和心得,便于自己查看,当然也想分享给走在前端路上的小伙伴(如果能帮到你一二,那也是极好的)。
小书中的每篇的篇幅都不是很长(单篇知识肯定没讲透),只是尽所能使其简单和让自己整明白各个模式是怎么一回事(如果也有让你整明白,那就更好了)。
来吧 =>
# 为什么要学设计(模式)
- 3 年工作经验,面试必考;
- 成为项目技术负责人,设计能力是必要基础;
- 从写好代码,到做好设计,设计模式是必经之路;
# 现实问题
- 网站资料针对 java 等后端语言比较多;
- 看懂概念,不知道怎么用,看完就忘;
- 现在的js框架(react、vue等),都用了哪些设计模式;
# 搭建开发环境
## 准备工作
```nodejs
# 项目初始化
npm init -y
# 新建开发目录src
mkdir src
# 安装webpack
npm install webpack-cli webpack --save-dev
# 安装babel
npm install babel-loader babel-core babel-preset-env html-webpack-plugin babel-plugin-transform-decorators-legacy -D
# 安装开发服务环境
npm install webpack-dev-server -D
# 新建配置webpack
touch webpack.dev.config.js
```
## 编写 `webpack.dev.config.js`
```javascript
const path = require('path')
const HtmlWebpackPlugin = require('html-webpack-plugin')
module.exports = {
entry: './src/index.js',
output: {
path: __dirname,
filename: './release/bundle.js' // release 会自动创建
},
plugins: [
new HtmlWebpackPlugin({
template: './index.html' // bundle.js 会自动注入
})
],
devServer: {
contentBase: path.join(__dirname, './release'), // 根目录
open: true, // 自动打开浏览器
port: 3000, // 端口
historyApiFallback: true
},
module: {
rules: [
{
test: /\.js?$/,
exclude: /(node_modules)/,
loader: 'babel-loader'
}
]
}
}
```
## 编辑 `package.json`,在 `script` 添加 `dev` 任务
```javascript
{
...,
"scripts": {
...,
"dev": "webpack --config ./webpack.dev.config.js --mode development"
},
...
}
```
# 面向对象
## 为什么使用面向对象
- 程序的执行离不开 **顺序、判断、循环** 操作,也就是将其结构化;
- 面向对象就是将零散的数据结构化;
- 对于计算机而言,结构化的才是最简单的(松本行弘的程序世界);
- 编程应该是**简单&抽象,简单的前提是抽象,抽象后才简单**;
> 关于抽象:抽取事物的共同特征就是抽取事物的本质特征,舍弃非本质的特征。所以抽象的过程也是一个裁剪的过程。在抽象时,同与不同,决定于从什么角度上来抽象。抽象的角度取决于分析问题的目的。
## 面向对象三要素
- 继承:子类继承父类(用的多);
- 封装:数据的权限和保密(将对象里面的某些属性和方法不想让别人看见。ES6 尚不支持,可用 typescript 演示);
- 多态:同一接口不同实现,简单来讲就是父类定义一个接口,子类实现不同的功能;
## 继承
```javascript
// 类,即模板
class People {
constructor(name, age) {
this.name = name
this.age = age
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
// 创建实例
const zhang = new People('张三', 27);
zhang.getName();
zhang.getAge();
// 创建实例
const li = new People('李四', 22);
zhang.getName();
zhang.getAge();
// 子类继承父类
class Student extends People{
constructor(name, age, id) {
super(name, age)
this.id = id
}
getId(){
alert(`${this.name},年龄 ${this.age},学号 ${this.id}`)
}
}
// 创建实例
const li = new Student('李四', 22, '001')
li.getId();
```
总结:
- `People` 是父类,公共的,不仅仅服务于 `Student`;
- 可将公共方法抽离出来,提高复用,减少冗余(这是软件设计最基础和最高效的方式);
## 封装
```javascript
// 封装 public-开放 protected-对子类开放 private-对自己开放
// 在线编译地址 => http://www.typescriptlang.org/play/
// 父类
class People{
public name
age
protected weight // 受保护属性,只有自己或子类可用
constructor(name, age) {
this.name = name
this.age = age
this.weight = 120
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
// 继承
class Student extends People {
id
private girlFriend
constructor(name, age, id) {
super(name, age)
this.id = id
this.girlFriend = '赵雪'
}
getId(){
alert(`${this.name},年龄 ${this.age},学号 ${this.id}`)
}
getWeight() {
alert(`${this.weight}`)
}
}
const xm = new Student('小明', 24, '002');
xm.getWeight();
// alert(xm.girlFriend)
```
说明
- `public` 完全开发;
- `portected` 对子类开放;
- `private` 对自己开放(ES6 尚不支持,可用 typescript 演示);
总结
- 较少耦合,不该外露的不外露
- 利于数据、接口的权限管理
- ES6目前不支持,一般认为 _开头的属性是 private ,比如var _num = 20
## 多态
```javascript
class A extends People {
constructor(name) {
super(name)
}
getName() {
alert(`A名字: ${this.name}`)
}
}
class B extends People {
constructor(name) {
super(name)
}
getName() {
alert(`B名字: ${this.name}`)
}
}
// 创建实例
const a1 = new A('张依依')
a1.getName();
// 创建实例
const b1 = new B('李尔尔')
b1.getName();
```
总结
- 保持子类的开放性和灵活性;
- 面向接口编程(不用管子类如何实现,就看父类有多少接口) ;
- js 应用极少;
- 需要结合 java 等语言的接口、重写、重载等功能;
# 应用举例
## 以 jQuery 为例
```javascript
class jQuery {
constructor(selector) {
let slice = Array.prototype.slice;
let dom = slice.call(document.querySelectorAll(selector));
let len = dom ? dom.length : 0;
for (let i = 0; i < len; i++) {
this[i] = dom[i];
}
this.length = len
this.selector = selector || ''
}
append(node) {
// ....
}
addClass(name) {
// ....
}
html(data) {
// ....
}
// 省略多个 API
}
window.$ = function(selector) {
// 工厂模式
return new jQuery(selector);
}
const $li = $('li')
console.log($li);
console.log($li.addClass);
```
# UML类图
UML,统一建模语言(Unified Modeling Language)。类图描述的是一种静态关系,在系统的整个生命周期都是有效的,是面向对象系统的建模中最常见的图,展现了一组对象、接口、协作和它们之间的关系。关系是指泛化(继承)和关联(引用)。
## 画图工具(工欲上其事必先利其器)
- MS Office visio;
- [processon](https://www.processon.com/)(不是会员有数量限制);
- Gliffy Digrams(chrome 应用插件,不过没有数量限制);
## 举例:一个简单类

```javascript
class People {
constructor(name, age) {
this.name = name
this.age = age
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
```
## 举例:继承与引用

```javascript
class House{
constructor(city) {
this.city = city;
}
showCity() {
alert(`城市:${this.city}`)
}
}
class People{
constructor(name, house) {
this.name = name
this.house = house
}
getInfo() {
alert(`我是${this.name},有房在【${this.house.city}】`)
}
}
class Student extends People{
constructor(name, house) {
super(name, house)
}
getInfo() {
alert(`我是${this.name},一名学生,有房在【${this.house.city}】`)
}
}
class Engineer extends People{
constructor(name, house) {
super(name, house)
}
getInfo() {
alert(`我是${this.name},一名工程师,有房在【${this.house.city}】`)
}
}
// 实例化
const h1 = new House('杭州');
const p1 = new People('张三', h1)
p1.getInfo();
const s1 = new Student('李四', h1)
s1.getInfo();
const e1 = new Engineer('王五', h1)
e1.getInfo();
```
# 你可以...
[下一篇:Javascript 设计模式之设计原则与 23 种设计模式](https://github.com/ruizhengyun/designPattern-note/issues/2) | 1.0 | Javascript 设计模式之面向对象与 UML 类图 - # 为什么要写 Javascript 设计模式小书
关于 Javascript 设计模式的文章与书有很多很多,我写这小书主要记录我的学习过程中的笔记和心得,便于自己查看,当然也想分享给走在前端路上的小伙伴(如果能帮到你一二,那也是极好的)。
小书中的每篇的篇幅都不是很长(单篇知识肯定没讲透),只是尽所能使其简单和让自己整明白各个模式是怎么一回事(如果也有让你整明白,那就更好了)。
来吧 =>
# 为什么要学设计(模式)
- 3 年工作经验,面试必考;
- 成为项目技术负责人,设计能力是必要基础;
- 从写好代码,到做好设计,设计模式是必经之路;
# 现实问题
- 网站资料针对 java 等后端语言比较多;
- 看懂概念,不知道怎么用,看完就忘;
- 现在的js框架(react、vue等),都用了哪些设计模式;
# 搭建开发环境
## 准备工作
```nodejs
# 项目初始化
npm init -y
# 新建开发目录src
mkdir src
# 安装webpack
npm install webpack-cli webpack --save-dev
# 安装babel
npm install babel-loader babel-core babel-preset-env html-webpack-plugin babel-plugin-transform-decorators-legacy -D
# 安装开发服务环境
npm install webpack-dev-server -D
# 新建配置webpack
touch webpack.dev.config.js
```
## 编写 `webpack.dev.config.js`
```javascript
const path = require('path')
const HtmlWebpackPlugin = require('html-webpack-plugin')
module.exports = {
entry: './src/index.js',
output: {
path: __dirname,
filename: './release/bundle.js' // release 会自动创建
},
plugins: [
new HtmlWebpackPlugin({
template: './index.html' // bundle.js 会自动注入
})
],
devServer: {
contentBase: path.join(__dirname, './release'), // 根目录
open: true, // 自动打开浏览器
port: 3000, // 端口
historyApiFallback: true
},
module: {
rules: [
{
test: /\.js?$/,
exclude: /(node_modules)/,
loader: 'babel-loader'
}
]
}
}
```
## 编辑 `package.json`,在 `script` 添加 `dev` 任务
```javascript
{
...,
"scripts": {
...,
"dev": "webpack --config ./webpack.dev.config.js --mode development"
},
...
}
```
# 面向对象
## 为什么使用面向对象
- 程序的执行离不开 **顺序、判断、循环** 操作,也就是将其结构化;
- 面向对象就是将零散的数据结构化;
- 对于计算机而言,结构化的才是最简单的(松本行弘的程序世界);
- 编程应该是**简单&抽象,简单的前提是抽象,抽象后才简单**;
> 关于抽象:抽取事物的共同特征就是抽取事物的本质特征,舍弃非本质的特征。所以抽象的过程也是一个裁剪的过程。在抽象时,同与不同,决定于从什么角度上来抽象。抽象的角度取决于分析问题的目的。
## 面向对象三要素
- 继承:子类继承父类(用的多);
- 封装:数据的权限和保密(将对象里面的某些属性和方法不想让别人看见。ES6 尚不支持,可用 typescript 演示);
- 多态:同一接口不同实现,简单来讲就是父类定义一个接口,子类实现不同的功能;
## 继承
```javascript
// 类,即模板
class People {
constructor(name, age) {
this.name = name
this.age = age
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
// 创建实例
const zhang = new People('张三', 27);
zhang.getName();
zhang.getAge();
// 创建实例
const li = new People('李四', 22);
zhang.getName();
zhang.getAge();
// 子类继承父类
class Student extends People{
constructor(name, age, id) {
super(name, age)
this.id = id
}
getId(){
alert(`${this.name},年龄 ${this.age},学号 ${this.id}`)
}
}
// 创建实例
const li = new Student('李四', 22, '001')
li.getId();
```
总结:
- `People` 是父类,公共的,不仅仅服务于 `Student`;
- 可将公共方法抽离出来,提高复用,减少冗余(这是软件设计最基础和最高效的方式);
## 封装
```javascript
// 封装 public-开放 protected-对子类开放 private-对自己开放
// 在线编译地址 => http://www.typescriptlang.org/play/
// 父类
class People{
public name
age
protected weight // 受保护属性,只有自己或子类可用
constructor(name, age) {
this.name = name
this.age = age
this.weight = 120
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
// 继承
class Student extends People {
id
private girlFriend
constructor(name, age, id) {
super(name, age)
this.id = id
this.girlFriend = '赵雪'
}
getId(){
alert(`${this.name},年龄 ${this.age},学号 ${this.id}`)
}
getWeight() {
alert(`${this.weight}`)
}
}
const xm = new Student('小明', 24, '002');
xm.getWeight();
// alert(xm.girlFriend)
```
说明
- `public` 完全开发;
- `portected` 对子类开放;
- `private` 对自己开放(ES6 尚不支持,可用 typescript 演示);
总结
- 较少耦合,不该外露的不外露
- 利于数据、接口的权限管理
- ES6目前不支持,一般认为 _开头的属性是 private ,比如var _num = 20
## 多态
```javascript
class A extends People {
constructor(name) {
super(name)
}
getName() {
alert(`A名字: ${this.name}`)
}
}
class B extends People {
constructor(name) {
super(name)
}
getName() {
alert(`B名字: ${this.name}`)
}
}
// 创建实例
const a1 = new A('张依依')
a1.getName();
// 创建实例
const b1 = new B('李尔尔')
b1.getName();
```
总结
- 保持子类的开放性和灵活性;
- 面向接口编程(不用管子类如何实现,就看父类有多少接口) ;
- js 应用极少;
- 需要结合 java 等语言的接口、重写、重载等功能;
# 应用举例
## 以 jQuery 为例
```javascript
class jQuery {
constructor(selector) {
let slice = Array.prototype.slice;
let dom = slice.call(document.querySelectorAll(selector));
let len = dom ? dom.length : 0;
for (let i = 0; i < len; i++) {
this[i] = dom[i];
}
this.length = len
this.selector = selector || ''
}
append(node) {
// ....
}
addClass(name) {
// ....
}
html(data) {
// ....
}
// 省略多个 API
}
window.$ = function(selector) {
// 工厂模式
return new jQuery(selector);
}
const $li = $('li')
console.log($li);
console.log($li.addClass);
```
# UML类图
UML,统一建模语言(Unified Modeling Language)。类图描述的是一种静态关系,在系统的整个生命周期都是有效的,是面向对象系统的建模中最常见的图,展现了一组对象、接口、协作和它们之间的关系。关系是指泛化(继承)和关联(引用)。
## 画图工具(工欲上其事必先利其器)
- MS Office visio;
- [processon](https://www.processon.com/)(不是会员有数量限制);
- Gliffy Digrams(chrome 应用插件,不过没有数量限制);
## 举例:一个简单类

```javascript
class People {
constructor(name, age) {
this.name = name
this.age = age
}
getName() {
alert(`名字: ${this.name}`)
}
getAge() {
alert(`年龄: ${this.age}`)
}
}
```
## 举例:继承与引用

```javascript
class House{
constructor(city) {
this.city = city;
}
showCity() {
alert(`城市:${this.city}`)
}
}
class People{
constructor(name, house) {
this.name = name
this.house = house
}
getInfo() {
alert(`我是${this.name},有房在【${this.house.city}】`)
}
}
class Student extends People{
constructor(name, house) {
super(name, house)
}
getInfo() {
alert(`我是${this.name},一名学生,有房在【${this.house.city}】`)
}
}
class Engineer extends People{
constructor(name, house) {
super(name, house)
}
getInfo() {
alert(`我是${this.name},一名工程师,有房在【${this.house.city}】`)
}
}
// 实例化
const h1 = new House('杭州');
const p1 = new People('张三', h1)
p1.getInfo();
const s1 = new Student('李四', h1)
s1.getInfo();
const e1 = new Engineer('王五', h1)
e1.getInfo();
```
# 你可以...
[下一篇:Javascript 设计模式之设计原则与 23 种设计模式](https://github.com/ruizhengyun/designPattern-note/issues/2) | non_priority | javascript 设计模式之面向对象与 uml 类图 为什么要写 javascript 设计模式小书 关于 javascript 设计模式的文章与书有很多很多,我写这小书主要记录我的学习过程中的笔记和心得,便于自己查看,当然也想分享给走在前端路上的小伙伴(如果能帮到你一二,那也是极好的)。 小书中的每篇的篇幅都不是很长(单篇知识肯定没讲透),只是尽所能使其简单和让自己整明白各个模式是怎么一回事(如果也有让你整明白,那就更好了)。 来吧 为什么要学设计(模式) 年工作经验,面试必考 成为项目技术负责人,设计能力是必要基础 从写好代码,到做好设计,设计模式是必经之路 现实问题 网站资料针对 java 等后端语言比较多 看懂概念,不知道怎么用,看完就忘 现在的js框架(react、vue等),都用了哪些设计模式 搭建开发环境 准备工作 nodejs 项目初始化 npm init y 新建开发目录src mkdir src 安装webpack npm install webpack cli webpack save dev 安装babel npm install babel loader babel core babel preset env html webpack plugin babel plugin transform decorators legacy d 安装开发服务环境 npm install webpack dev server d 新建配置webpack touch webpack dev config js 编写 webpack dev config js javascript const path require path const htmlwebpackplugin require html webpack plugin module exports entry src index js output path dirname filename release bundle js release 会自动创建 plugins new htmlwebpackplugin template index html bundle js 会自动注入 devserver contentbase path join dirname release 根目录 open true 自动打开浏览器 port 端口 historyapifallback true module rules test js exclude node modules loader babel loader 编辑 package json ,在 script 添加 dev 任务 javascript scripts dev webpack config webpack dev config js mode development 面向对象 为什么使用面向对象 程序的执行离不开 顺序、判断、循环 操作,也就是将其结构化 面向对象就是将零散的数据结构化 对于计算机而言,结构化的才是最简单的(松本行弘的程序世界) 编程应该是 简单 抽象,简单的前提是抽象,抽象后才简单 关于抽象:抽取事物的共同特征就是抽取事物的本质特征,舍弃非本质的特征。所以抽象的过程也是一个裁剪的过程。在抽象时,同与不同,决定于从什么角度上来抽象。抽象的角度取决于分析问题的目的。 面向对象三要素 继承:子类继承父类(用的多) 封装:数据的权限和保密(将对象里面的某些属性和方法不想让别人看见。 尚不支持,可用 typescript 演示) 多态:同一接口不同实现,简单来讲就是父类定义一个接口,子类实现不同的功能 继承 javascript 类,即模板 class people constructor name age this name name this age age getname alert 名字 this name getage alert 年龄 this age 创建实例 const zhang new people 张三 zhang getname zhang getage 创建实例 const li new people 李四 zhang getname zhang getage 子类继承父类 class student extends people constructor name age id super name age this id id getid alert this name ,年龄 this age ,学号 this id 创建实例 const li new student 李四 li getid 总结: people 是父类,公共的,不仅仅服务于 student 可将公共方法抽离出来,提高复用,减少冗余(这是软件设计最基础和最高效的方式) 封装 javascript 封装 public 开放 protected 对子类开放 private 对自己开放 在线编译地址 父类 class people public name age protected weight 受保护属性,只有自己或子类可用 constructor name age this name name this age age this weight getname alert 名字 this name getage alert 年龄 this age 继承 class student extends people id private girlfriend constructor name age id super name age this id id this girlfriend 赵雪 getid alert this name ,年龄 this age ,学号 this id getweight alert this weight const xm new student 小明 xm getweight alert xm girlfriend 说明 public 完全开发; portected 对子类开放; private 对自己开放 尚不支持,可用 typescript 演示 ; 总结 较少耦合,不该外露的不外露 利于数据、接口的权限管理 ,一般认为 开头的属性是 private ,比如var num 多态 javascript class a extends people constructor name super name getname alert a名字 this name class b extends people constructor name super name getname alert b名字 this name 创建实例 const new a 张依依 getname 创建实例 const new b 李尔尔 getname 总结 保持子类的开放性和灵活性 面向接口编程 不用管子类如何实现,就看父类有多少接口 js 应用极少 需要结合 java 等语言的接口、重写、重载等功能 应用举例 以 jquery 为例 javascript class jquery constructor selector let slice array prototype slice let dom slice call document queryselectorall selector let len dom dom length for let i i len i this dom this length len this selector selector append node addclass name html data 省略多个 api window function selector 工厂模式 return new jquery selector const li li console log li console log li addclass uml类图 uml,统一建模语言(unified modeling language)。类图描述的是一种静态关系,在系统的整个生命周期都是有效的,是面向对象系统的建模中最常见的图,展现了一组对象、接口、协作和它们之间的关系。关系是指泛化(继承)和关联(引用)。 画图工具(工欲上其事必先利其器) ms office visio gliffy digrams(chrome 应用插件,不过没有数量限制) 举例:一个简单类 javascript class people constructor name age this name name this age age getname alert 名字 this name getage alert 年龄 this age 举例:继承与引用 javascript class house constructor city this city city showcity alert 城市: this city class people constructor name house this name name this house house getinfo alert 我是 this name ,有房在【 this house city 】 class student extends people constructor name house super name house getinfo alert 我是 this name ,一名学生,有房在【 this house city 】 class engineer extends people constructor name house super name house getinfo alert 我是 this name ,一名工程师,有房在【 this house city 】 实例化 const new house 杭州 const new people 张三 getinfo const new student 李四 getinfo const new engineer 王五 getinfo 你可以 | 0 |
10,647 | 4,075,420,155 | IssuesEvent | 2016-05-29 06:45:27 | DotSpatial/DotSpatial | https://api.github.com/repos/DotSpatial/DotSpatial | closed | Cannot delete newly added features | bug CodePlex | **This issue was imported from [CodePlex](http://dotspatial.codeplex.com/workitem/175)**
**[keenedge](http://www.codeplex.com/site/users/view/keenedge)** wrote 2010-11-30 at 02:41
With DemoMap
<ul>
<li>Add a layer </li><li>Enable the ShapeEditor tool </li><li>Digitize a shape </li><li>Using the select tool select the newly digitized shape. </li><li>Use the "remove Slected features" tool to remove the shape.
</li></ul>
Nothing happens. The new feature is not removed.
**[keenedge](http://www.codeplex.com/site/users/view/keenedge)** wrote 2010-12-06 at 21:30
Anyone have anything thoughts on why this is? It is a result of one of the recent change sets.
**[Shade1974](http://www.codeplex.com/site/users/view/Shade1974)** wrote 2010-12-06 at 23:02
Er.. it is obviously a bug. It might have to do with changing from index mode to Feature mode, but it's not intuitive to me as to what might be wrong since the vertices have to be defined correctly or else it couldn't draw the feature. I wonder if this
is specific to one feature type or if this is true with all feature types. I'm sure I can get to the bottom of it, I just have been working on getting our documentation engine online. I'll be back to debugging probably later this week though and can look at
testing it then.
Ted
**[geofran80](http://www.codeplex.com/site/users/view/geofran80)** wrote 2011-03-31 at 15:20
Dear friends, is fixed this issue?
Thanks.
Francisco J. | 1.0 | Cannot delete newly added features - **This issue was imported from [CodePlex](http://dotspatial.codeplex.com/workitem/175)**
**[keenedge](http://www.codeplex.com/site/users/view/keenedge)** wrote 2010-11-30 at 02:41
With DemoMap
<ul>
<li>Add a layer </li><li>Enable the ShapeEditor tool </li><li>Digitize a shape </li><li>Using the select tool select the newly digitized shape. </li><li>Use the "remove Slected features" tool to remove the shape.
</li></ul>
Nothing happens. The new feature is not removed.
**[keenedge](http://www.codeplex.com/site/users/view/keenedge)** wrote 2010-12-06 at 21:30
Anyone have anything thoughts on why this is? It is a result of one of the recent change sets.
**[Shade1974](http://www.codeplex.com/site/users/view/Shade1974)** wrote 2010-12-06 at 23:02
Er.. it is obviously a bug. It might have to do with changing from index mode to Feature mode, but it's not intuitive to me as to what might be wrong since the vertices have to be defined correctly or else it couldn't draw the feature. I wonder if this
is specific to one feature type or if this is true with all feature types. I'm sure I can get to the bottom of it, I just have been working on getting our documentation engine online. I'll be back to debugging probably later this week though and can look at
testing it then.
Ted
**[geofran80](http://www.codeplex.com/site/users/view/geofran80)** wrote 2011-03-31 at 15:20
Dear friends, is fixed this issue?
Thanks.
Francisco J. | non_priority | cannot delete newly added features this issue was imported from wrote at with demomap add a layer enable the shapeeditor tool digitize a shape using the select tool select the newly digitized shape use the remove slected features tool to remove the shape nothing happens the new feature is not removed wrote at anyone have anything thoughts on why this is it is a result of one of the recent change sets wrote at er it is obviously a bug it might have to do with changing from index mode to feature mode but it s not intuitive to me as to what might be wrong since the vertices have to be defined correctly or else it couldn t draw the feature i wonder if this is specific to one feature type or if this is true with all feature types i m sure i can get to the bottom of it i just have been working on getting our documentation engine online i ll be back to debugging probably later this week though and can look at testing it then ted wrote at dear friends is fixed this issue thanks francisco j | 0 |
110,182 | 11,693,715,268 | IssuesEvent | 2020-03-06 01:28:53 | Seneca-CDOT/telescope | https://api.github.com/repos/Seneca-CDOT/telescope | closed | Documentation should reflect redis conflict when running it locally and with docker | area: redis type: bug type: documentation | Our documentation should mention that the methods suggested to run redis can't be used at the same time and, in case of redis error, the developer should check if another instance of redis is running using any other method.
Example:
Developer tries `docker-compose up redis` and gets:
> ERROR: for telescope_redis_1 Cannot start service redis: driver failed programming external connectivity on endpoint telescope_redis_1 (335448dd9bed3b7bd79109979faa0100288e3a8abe8366992600cf22ad382adf): Error starting userland proxy: listen tcp 0.0.0.0:6379: bind: address already in use
The problem is that redis is already running natively so it should be stopped (CLI and services should be checked).
| 1.0 | Documentation should reflect redis conflict when running it locally and with docker - Our documentation should mention that the methods suggested to run redis can't be used at the same time and, in case of redis error, the developer should check if another instance of redis is running using any other method.
Example:
Developer tries `docker-compose up redis` and gets:
> ERROR: for telescope_redis_1 Cannot start service redis: driver failed programming external connectivity on endpoint telescope_redis_1 (335448dd9bed3b7bd79109979faa0100288e3a8abe8366992600cf22ad382adf): Error starting userland proxy: listen tcp 0.0.0.0:6379: bind: address already in use
The problem is that redis is already running natively so it should be stopped (CLI and services should be checked).
| non_priority | documentation should reflect redis conflict when running it locally and with docker our documentation should mention that the methods suggested to run redis can t be used at the same time and in case of redis error the developer should check if another instance of redis is running using any other method example developer tries docker compose up redis and gets error for telescope redis cannot start service redis driver failed programming external connectivity on endpoint telescope redis error starting userland proxy listen tcp bind address already in use the problem is that redis is already running natively so it should be stopped cli and services should be checked | 0 |
42,321 | 5,437,789,002 | IssuesEvent | 2017-03-06 08:27:47 | geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE | https://api.github.com/repos/geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE | reopened | idd41QvOU+Ju6iidqpYGQMifQvh2LeGtWRpaYHIBKIOEyWv4xrOuOjX58K+ECjLeL2/+a42Ts0VwEeCxcbN4/YWZb4XnBu183Fn+2DfwRcNJ6p+lB6bQBu2Ck8lIZTwwvGosrPRrjd2NmGyXMBGkiXmigdYYX1Aw6QljWx63LHk= | design | scJJNQ6eM8wHDb5SpbHb4KQTniSEjrCjf3LGOpORAYVtIMWE6pYfUWC9n8SjkbzOoIG74D04+lXCJ7uuo+RxcSD+FSgvW125BOiyLsU12j2+TI1XbgHijO5XlIL1KdKpzvQpqy+IWNn2h5tp7xiLYZs/JkUbc2MSshyVkEmPCjj0eQ/3dbYW5VBrlMnBClsH8sHnz2anPQzxcQm8ElvOfeikSEwiGmQdR7DKKag02EwVu06ve863jWnZ8BXZfTiL8sHnz2anPQzxcQm8ElvOfRuEzy44G8bG8xPkSQLOt67UndbQZXx9NQSeeWTVW6NN6qqmhs/8D4YwuM4HHnXxN/LB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW859iEbhUIc4UrUlanaE8T+3dEAGgzIqvYJaIF5VOGAcBhDywefPZqc9DPFxCbwSW8598HIwfVeqPYhApCsxG8C0Xo3iIwbt/I2I/X6e3JTxPxPywefPZqc9DPFxCbwSW859RylcIqEscXuQ4RKuCYSWNaSC1z46o2KKnzcJm420wFjKY3dP5hqyIlFgBYUwbN6E8GyZjw2O79/IjxbS2lBOBfLB589mpz0M8XEJvBJbzn3UdunS/WlbAT+VQj9TitoM1+8MwAgt/3AZq8S60VykC7bg8a19fTyOKmlme/GQSfvywefPZqc9DPFxCbwSW859A/JnFXL36CAU28sgwBNlIHr+53YdJkVhxXSFGLqEoZQ2zaw4aKvCjmNo7G74eITF9g1vHAAkG6Qfr/Kuohxi0g+IV/x82Mesc1/+1FhhE0gYONHnHPcbwLjcR0jpAnLcLi6fD/5xv3wK4IxzkBRxUY0j0D6kGMFFGIst8OoVUpLywefPZqc9DPFxCbwSW859Ap/TVJ/Qc7LiwOhQzUMILzS1B6hY7t6wkpg4noyB+o/cvXqv3mTMFrXewgAJs9zMhNdm/ij/4LU2QNXbtoP/tlMF4a+depb10X7d6kxcqWK76+/PxeJMhjURpmkzyr3oChXD7O5cEynonhO3WLVtjfLB589mpz0M8XEJvBJbzn3sxDsLBhh28JKpbOvgNco9PVOdsVRQD++eOLsT90HBW/l+E2n4v70u6xBHNfT5/P0DToZP3bozC14tGLPQnx7tCwp26HuEiSjzMbwuUl6xu7HfHEalkfALh8rNI8HRQW/ywefPZqc9DPFxCbwSW8596TVM/sYhVJETuhPO3CzgELM9mzAQ2nfzkJBY1rIM8SCeFQk3ciBT7bpV2vou8LUf8HfholLZu3J5xf/iheZVPjhcrLW44wq02U7Xb7qqeZ5LuuR+MXsSj94HahN9ouWPLvWMCt9MMuDGLCL7tktmJvn66IsDjfEb01kO5SlEqL6dKHGJIzyc4V76LjahyBWg8sHnz2anPQzxcQm8ElvOfcxjhUjHeSFtf0314dFcXBZnwGyDsILV5XbqsoxBDEgv8sHnz2anPQzxcQm8ElvOfXVQihMNSsx5K02f+sa7J1GwY6ooDq0zhmcHl/BnMbp4eSvtg9JD9AwdK5eQZLbOlVpPrSiGL1tlFMuR0hxORKztcIcMQTssGcQPYKOAgiHQVEjdT/hu8vdJuAnKPrOzjpsPWzdKjhSRaEfmRc+obNYabf789/USkrCdMVf/EIuBGRIVQZaTJZIYQvzcfucZIO4APy/AmnDl3MN/htzPdVGU5iy14j91mD1Ak6xKbE6fJicHSqGVv0RPlqXGndxwWyAPdlcdB/K/J1pUja/yJBHgsi29eFiy903fmycDI+kJ8sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn3zKrV3XqofJrsU/4FK9e0T6sc3RtTs31cGvQq7OYjtMPLB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW8599RSZiJprDshBA75L5ZKTpk2or6J/Y222SjDpo8Z/ybcEC9ZKH0fJazw9wuFh5nq58sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn3BncqH38mGz9PVkDRqCWKbhn3ibZ3/el3OIXLkwyw0TfLB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW859lJHDQcZUdfTYTf8ZHrodLe9B+qUXUYpBJ86mNMZYuvjCULkTWl5HysIdyqTRINp08sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn0X49q801d7umXSphX5uyJzuvvm0iJ0CYaIioJrcvNyRzxDlv8dHPUSx96sBw73hV8IAUzb4MO8O3orSJe0QJ1qmOW9EJGtK1h1YQzrrBLj1hyYWBNmP4iDZsDoJABSxxaGDqQ0l/s7ytodJCpOIAzZfBLrbHCfkbiBwHNkiwtman0QcJOuUqdHxRMUCa7M+W8= | 1.0 | idd41QvOU+Ju6iidqpYGQMifQvh2LeGtWRpaYHIBKIOEyWv4xrOuOjX58K+ECjLeL2/+a42Ts0VwEeCxcbN4/YWZb4XnBu183Fn+2DfwRcNJ6p+lB6bQBu2Ck8lIZTwwvGosrPRrjd2NmGyXMBGkiXmigdYYX1Aw6QljWx63LHk= - scJJNQ6eM8wHDb5SpbHb4KQTniSEjrCjf3LGOpORAYVtIMWE6pYfUWC9n8SjkbzOoIG74D04+lXCJ7uuo+RxcSD+FSgvW125BOiyLsU12j2+TI1XbgHijO5XlIL1KdKpzvQpqy+IWNn2h5tp7xiLYZs/JkUbc2MSshyVkEmPCjj0eQ/3dbYW5VBrlMnBClsH8sHnz2anPQzxcQm8ElvOfeikSEwiGmQdR7DKKag02EwVu06ve863jWnZ8BXZfTiL8sHnz2anPQzxcQm8ElvOfRuEzy44G8bG8xPkSQLOt67UndbQZXx9NQSeeWTVW6NN6qqmhs/8D4YwuM4HHnXxN/LB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW859iEbhUIc4UrUlanaE8T+3dEAGgzIqvYJaIF5VOGAcBhDywefPZqc9DPFxCbwSW8598HIwfVeqPYhApCsxG8C0Xo3iIwbt/I2I/X6e3JTxPxPywefPZqc9DPFxCbwSW859RylcIqEscXuQ4RKuCYSWNaSC1z46o2KKnzcJm420wFjKY3dP5hqyIlFgBYUwbN6E8GyZjw2O79/IjxbS2lBOBfLB589mpz0M8XEJvBJbzn3UdunS/WlbAT+VQj9TitoM1+8MwAgt/3AZq8S60VykC7bg8a19fTyOKmlme/GQSfvywefPZqc9DPFxCbwSW859A/JnFXL36CAU28sgwBNlIHr+53YdJkVhxXSFGLqEoZQ2zaw4aKvCjmNo7G74eITF9g1vHAAkG6Qfr/Kuohxi0g+IV/x82Mesc1/+1FhhE0gYONHnHPcbwLjcR0jpAnLcLi6fD/5xv3wK4IxzkBRxUY0j0D6kGMFFGIst8OoVUpLywefPZqc9DPFxCbwSW859Ap/TVJ/Qc7LiwOhQzUMILzS1B6hY7t6wkpg4noyB+o/cvXqv3mTMFrXewgAJs9zMhNdm/ij/4LU2QNXbtoP/tlMF4a+depb10X7d6kxcqWK76+/PxeJMhjURpmkzyr3oChXD7O5cEynonhO3WLVtjfLB589mpz0M8XEJvBJbzn3sxDsLBhh28JKpbOvgNco9PVOdsVRQD++eOLsT90HBW/l+E2n4v70u6xBHNfT5/P0DToZP3bozC14tGLPQnx7tCwp26HuEiSjzMbwuUl6xu7HfHEalkfALh8rNI8HRQW/ywefPZqc9DPFxCbwSW8596TVM/sYhVJETuhPO3CzgELM9mzAQ2nfzkJBY1rIM8SCeFQk3ciBT7bpV2vou8LUf8HfholLZu3J5xf/iheZVPjhcrLW44wq02U7Xb7qqeZ5LuuR+MXsSj94HahN9ouWPLvWMCt9MMuDGLCL7tktmJvn66IsDjfEb01kO5SlEqL6dKHGJIzyc4V76LjahyBWg8sHnz2anPQzxcQm8ElvOfcxjhUjHeSFtf0314dFcXBZnwGyDsILV5XbqsoxBDEgv8sHnz2anPQzxcQm8ElvOfXVQihMNSsx5K02f+sa7J1GwY6ooDq0zhmcHl/BnMbp4eSvtg9JD9AwdK5eQZLbOlVpPrSiGL1tlFMuR0hxORKztcIcMQTssGcQPYKOAgiHQVEjdT/hu8vdJuAnKPrOzjpsPWzdKjhSRaEfmRc+obNYabf789/USkrCdMVf/EIuBGRIVQZaTJZIYQvzcfucZIO4APy/AmnDl3MN/htzPdVGU5iy14j91mD1Ak6xKbE6fJicHSqGVv0RPlqXGndxwWyAPdlcdB/K/J1pUja/yJBHgsi29eFiy903fmycDI+kJ8sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn3zKrV3XqofJrsU/4FK9e0T6sc3RtTs31cGvQq7OYjtMPLB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW8599RSZiJprDshBA75L5ZKTpk2or6J/Y222SjDpo8Z/ybcEC9ZKH0fJazw9wuFh5nq58sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn3BncqH38mGz9PVkDRqCWKbhn3ibZ3/el3OIXLkwyw0TfLB589mpz0M8XEJvBJbzn3ywefPZqc9DPFxCbwSW859lJHDQcZUdfTYTf8ZHrodLe9B+qUXUYpBJ86mNMZYuvjCULkTWl5HysIdyqTRINp08sHnz2anPQzxcQm8ElvOffLB589mpz0M8XEJvBJbzn0X49q801d7umXSphX5uyJzuvvm0iJ0CYaIioJrcvNyRzxDlv8dHPUSx96sBw73hV8IAUzb4MO8O3orSJe0QJ1qmOW9EJGtK1h1YQzrrBLj1hyYWBNmP4iDZsDoJABSxxaGDqQ0l/s7ytodJCpOIAzZfBLrbHCfkbiBwHNkiwtman0QcJOuUqdHxRMUCa7M+W8= | non_priority | rxcsd wlbat iv tvj o ij l uskrcdmvf k | 0 |
266,593 | 20,158,809,582 | IssuesEvent | 2022-02-09 19:08:49 | kcleal/dysgu | https://api.github.com/repos/kcleal/dysgu | reopened | Installation Note | documentation | I was interested in trying Dysgu, and created a fresh conda environment with python 3.10. When I tried to install dysgu within that fresh env, I got the following error:
```
(dysgu)$ pip install dysgu
ERROR: Could not find a version that satisfies the requirement dysgu (from versions: none)
ERROR: No matching distribution found for dysgu
```
This seemed odd to me, and may have something to do with it the new version released 1 day before.
However, when I dropped my python version down to 3.7 it went through (I did have to install `numpy` first). This is just for your documentation or others' troubleshooting.
| 1.0 | Installation Note - I was interested in trying Dysgu, and created a fresh conda environment with python 3.10. When I tried to install dysgu within that fresh env, I got the following error:
```
(dysgu)$ pip install dysgu
ERROR: Could not find a version that satisfies the requirement dysgu (from versions: none)
ERROR: No matching distribution found for dysgu
```
This seemed odd to me, and may have something to do with it the new version released 1 day before.
However, when I dropped my python version down to 3.7 it went through (I did have to install `numpy` first). This is just for your documentation or others' troubleshooting.
| non_priority | installation note i was interested in trying dysgu and created a fresh conda environment with python when i tried to install dysgu within that fresh env i got the following error dysgu pip install dysgu error could not find a version that satisfies the requirement dysgu from versions none error no matching distribution found for dysgu this seemed odd to me and may have something to do with it the new version released day before however when i dropped my python version down to it went through i did have to install numpy first this is just for your documentation or others troubleshooting | 0 |
334,938 | 29,999,206,828 | IssuesEvent | 2023-06-26 08:10:02 | ITISFoundation/osparc-ops-environments | https://api.github.com/repos/ITISFoundation/osparc-ops-environments | opened | e2e-ops: Remove hardcoded credentials. | p:low-prio FAST e2e & test | Switch hardcoded credentials to access tokens for pgbackup and mail-server e2e-ops-test. Some login credentials to github are hardcoded. We can switch this to access tokens for better security and manageability. This is not urgent/critical. | 1.0 | e2e-ops: Remove hardcoded credentials. - Switch hardcoded credentials to access tokens for pgbackup and mail-server e2e-ops-test. Some login credentials to github are hardcoded. We can switch this to access tokens for better security and manageability. This is not urgent/critical. | non_priority | ops remove hardcoded credentials switch hardcoded credentials to access tokens for pgbackup and mail server ops test some login credentials to github are hardcoded we can switch this to access tokens for better security and manageability this is not urgent critical | 0 |
6,900 | 3,928,351,745 | IssuesEvent | 2016-04-24 06:36:51 | haskell/cabal | https://api.github.com/repos/haskell/cabal | closed | [nix-local-build] unbounded-delays Setup on GHC 8.0rc2 causing problems for nix-local-build | nix-local-build | When I attempt to new-build the nix-local-branch with GHC 8.0rc2 I get the following error:
```
ezyang@sabre:~/Dev/cabal-nix-local-build$ cabal new-build -w ghc-8.0 --enable-tests
Resolving dependencies...
cabal: Could not resolve dependencies:
trying: Cabal-test.tasty-quickcheck~>cabal-install-test.tasty-quickcheck-0.8.4
(dependency of cabal-install-test.Cabal-1.24.0.0:*test)
trying: Cabal-test.tasty-quickcheck-0.8.4:!test
trying: Cabal-test.tasty-hunit~>cabal-install-test.tasty-hunit-0.9.2
(dependency of cabal-install-test.Cabal-1.24.0.0:*test)
trying: Cabal-test.tasty~>cabal-install-test.tasty-0.11.0.2 (dependency of
cabal-install-test.Cabal-1.24.0.0:*test)
trying:
Cabal-test.unbounded-delays~>cabal-install-test.unbounded-delays-0.1.0.9
(dependency of Cabal-test.tasty-0.11.0.2)
trying: Cabal-test.tasty-quickcheck-0.8.4:-old-quickcheck
trying: unbounded-delays-setup.Cabal~>cabal-install-test.Cabal-1.24.0.0
(dependency of cabal-install-test.unbounded-delays-0.1.0.9)
rejecting: unbounded-delays-setup.Cabal-1.24.0.0:!test (constraint from config
file, command line flag, or user target requires opposite flag selection)
trying: unbounded-delays-setup.Cabal-1.24.0.0:*test
trying: unbounded-delays-setup.Cabal-1.24.0.0:-bundled-binary-generic
Backjump limit reached (currently 2000, change with --max-backjumps or try to
run with --reorder-goals).
```
This doesn't happen if I use GHC 7.10. I don't really know how to read this error message, which is another problem in and of itself.
I can "solve" the problem by vendoring unbounded-delays to have a non-Custom setup script. | 1.0 | [nix-local-build] unbounded-delays Setup on GHC 8.0rc2 causing problems for nix-local-build - When I attempt to new-build the nix-local-branch with GHC 8.0rc2 I get the following error:
```
ezyang@sabre:~/Dev/cabal-nix-local-build$ cabal new-build -w ghc-8.0 --enable-tests
Resolving dependencies...
cabal: Could not resolve dependencies:
trying: Cabal-test.tasty-quickcheck~>cabal-install-test.tasty-quickcheck-0.8.4
(dependency of cabal-install-test.Cabal-1.24.0.0:*test)
trying: Cabal-test.tasty-quickcheck-0.8.4:!test
trying: Cabal-test.tasty-hunit~>cabal-install-test.tasty-hunit-0.9.2
(dependency of cabal-install-test.Cabal-1.24.0.0:*test)
trying: Cabal-test.tasty~>cabal-install-test.tasty-0.11.0.2 (dependency of
cabal-install-test.Cabal-1.24.0.0:*test)
trying:
Cabal-test.unbounded-delays~>cabal-install-test.unbounded-delays-0.1.0.9
(dependency of Cabal-test.tasty-0.11.0.2)
trying: Cabal-test.tasty-quickcheck-0.8.4:-old-quickcheck
trying: unbounded-delays-setup.Cabal~>cabal-install-test.Cabal-1.24.0.0
(dependency of cabal-install-test.unbounded-delays-0.1.0.9)
rejecting: unbounded-delays-setup.Cabal-1.24.0.0:!test (constraint from config
file, command line flag, or user target requires opposite flag selection)
trying: unbounded-delays-setup.Cabal-1.24.0.0:*test
trying: unbounded-delays-setup.Cabal-1.24.0.0:-bundled-binary-generic
Backjump limit reached (currently 2000, change with --max-backjumps or try to
run with --reorder-goals).
```
This doesn't happen if I use GHC 7.10. I don't really know how to read this error message, which is another problem in and of itself.
I can "solve" the problem by vendoring unbounded-delays to have a non-Custom setup script. | non_priority | unbounded delays setup on ghc causing problems for nix local build when i attempt to new build the nix local branch with ghc i get the following error ezyang sabre dev cabal nix local build cabal new build w ghc enable tests resolving dependencies cabal could not resolve dependencies trying cabal test tasty quickcheck cabal install test tasty quickcheck dependency of cabal install test cabal test trying cabal test tasty quickcheck test trying cabal test tasty hunit cabal install test tasty hunit dependency of cabal install test cabal test trying cabal test tasty cabal install test tasty dependency of cabal install test cabal test trying cabal test unbounded delays cabal install test unbounded delays dependency of cabal test tasty trying cabal test tasty quickcheck old quickcheck trying unbounded delays setup cabal cabal install test cabal dependency of cabal install test unbounded delays rejecting unbounded delays setup cabal test constraint from config file command line flag or user target requires opposite flag selection trying unbounded delays setup cabal test trying unbounded delays setup cabal bundled binary generic backjump limit reached currently change with max backjumps or try to run with reorder goals this doesn t happen if i use ghc i don t really know how to read this error message which is another problem in and of itself i can solve the problem by vendoring unbounded delays to have a non custom setup script | 0 |
56,893 | 6,532,878,235 | IssuesEvent | 2017-08-31 02:11:02 | EFForg/https-everywhere | https://api.github.com/repos/EFForg/https-everywhere | closed | hsts-prune is still pointed to the discontinued Firefox Aurora release | bug Ruleset Testing | The preload list for Firefox Aurora [hasn't been updated since April](https://hg.mozilla.org/releases/mozilla-aurora/log/tip/security/manager/ssl/nsSTSPreloadList.inc). We'll need to point hsts-prune to Firefox Nightly instead, and make changes to support the new format of `nsSTSPreloadList.inc`. | 1.0 | hsts-prune is still pointed to the discontinued Firefox Aurora release - The preload list for Firefox Aurora [hasn't been updated since April](https://hg.mozilla.org/releases/mozilla-aurora/log/tip/security/manager/ssl/nsSTSPreloadList.inc). We'll need to point hsts-prune to Firefox Nightly instead, and make changes to support the new format of `nsSTSPreloadList.inc`. | non_priority | hsts prune is still pointed to the discontinued firefox aurora release the preload list for firefox aurora we ll need to point hsts prune to firefox nightly instead and make changes to support the new format of nsstspreloadlist inc | 0 |
117,543 | 11,949,544,988 | IssuesEvent | 2020-04-03 13:49:39 | abdessamed11/C2N3_C3N3 | https://api.github.com/repos/abdessamed11/C2N3_C3N3 | closed | Préparation de l'environnement | documentation | - [x] FORK
- [x] npm install /packages/
- [x] Watching concatenation Préfixage avec NPM Scripts
- [x] Compression du CSS et Javascript avec NPM Scripts | 1.0 | Préparation de l'environnement - - [x] FORK
- [x] npm install /packages/
- [x] Watching concatenation Préfixage avec NPM Scripts
- [x] Compression du CSS et Javascript avec NPM Scripts | non_priority | préparation de l environnement fork npm install packages watching concatenation préfixage avec npm scripts compression du css et javascript avec npm scripts | 0 |
96,586 | 12,140,998,267 | IssuesEvent | 2020-04-23 21:32:43 | NCEAS/metacatui | https://api.github.com/repos/NCEAS/metacatui | opened | Add "new features" widget | design documentation enhancement | Design and implement a non-obtrusive but noticable widget for MetacatUI that will show what the new features of the latest release are. We can directly use the Release Notes from here on Github, which I usually try to write in a non-jargony way as much as possible. | 1.0 | Add "new features" widget - Design and implement a non-obtrusive but noticable widget for MetacatUI that will show what the new features of the latest release are. We can directly use the Release Notes from here on Github, which I usually try to write in a non-jargony way as much as possible. | non_priority | add new features widget design and implement a non obtrusive but noticable widget for metacatui that will show what the new features of the latest release are we can directly use the release notes from here on github which i usually try to write in a non jargony way as much as possible | 0 |
215,393 | 16,670,097,083 | IssuesEvent | 2021-06-07 09:45:59 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | No response after clicking 'New Window' on Storage Explorer which is installed by snap | :computer: linux 🧪 testing | **Storage Explorer Version**: 1.20.0-dev
**Build Number**: 20210605.1
**Branch**: main
**Platform/OS**: Linux Ubuntu (18.04/ 16.04/ 20.04)/ CentOS 7.6.1810
**Architecture**: ix64
**How Found**: From running test case
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Install and launch Storage Explorer on Linux by snap.
2. Click 'File -> New Window'.
3. Check whether a new window opens.
## Expected Experience ##
A new window opens.
## Actual Experience ##
No response after clicking 'New Window'.
## Additional Context ##
This issue doesn't reproduce on storage explorer which is installed by using .tar file. | 1.0 | No response after clicking 'New Window' on Storage Explorer which is installed by snap - **Storage Explorer Version**: 1.20.0-dev
**Build Number**: 20210605.1
**Branch**: main
**Platform/OS**: Linux Ubuntu (18.04/ 16.04/ 20.04)/ CentOS 7.6.1810
**Architecture**: ix64
**How Found**: From running test case
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Install and launch Storage Explorer on Linux by snap.
2. Click 'File -> New Window'.
3. Check whether a new window opens.
## Expected Experience ##
A new window opens.
## Actual Experience ##
No response after clicking 'New Window'.
## Additional Context ##
This issue doesn't reproduce on storage explorer which is installed by using .tar file. | non_priority | no response after clicking new window on storage explorer which is installed by snap storage explorer version dev build number branch main platform os linux ubuntu centos architecture how found from running test case regression from not a regression steps to reproduce install and launch storage explorer on linux by snap click file new window check whether a new window opens expected experience a new window opens actual experience no response after clicking new window additional context this issue doesn t reproduce on storage explorer which is installed by using tar file | 0 |
129,872 | 17,929,654,416 | IssuesEvent | 2021-09-10 07:29:30 | decidim/decidim | https://api.github.com/repos/decidim/decidim | closed | Sign the digital electoral closure certificate | status: design required contract: e-voting module: elections | ref V-PGMF.08
As the Polling Station Officer, I want to be able to sign and validate that the data that I've uploaded to the web system (the digital electoral closure certificate) and the one that I have in the physical closure certificate are the same.
I should have a checkbox or some way that says: "I've reviewed this and is the same as the physical electoral closure certificate".
This is closely related to:
* Fill in a web form the results (digital electoral closure certificate) (ref V-PGMF.09))
* Capture a picture of the physical electoral closure certificate (ref V-PGMF.11)
As it's the confirmation that both of these certificates have the same information.
After I've confirmed it, I can no longer modify any of the information, so I need to have a confirmation dialog explaining that this action can't be undone.
### Acceptance criteria
- [ ] As a Polling Station Officer I can confirm that the physical and digital closure certificate have the same information
- [ ] I see a confirmation dialog when making this action
- [ ] After I confirmed I can't undo anything
| 1.0 | Sign the digital electoral closure certificate - ref V-PGMF.08
As the Polling Station Officer, I want to be able to sign and validate that the data that I've uploaded to the web system (the digital electoral closure certificate) and the one that I have in the physical closure certificate are the same.
I should have a checkbox or some way that says: "I've reviewed this and is the same as the physical electoral closure certificate".
This is closely related to:
* Fill in a web form the results (digital electoral closure certificate) (ref V-PGMF.09))
* Capture a picture of the physical electoral closure certificate (ref V-PGMF.11)
As it's the confirmation that both of these certificates have the same information.
After I've confirmed it, I can no longer modify any of the information, so I need to have a confirmation dialog explaining that this action can't be undone.
### Acceptance criteria
- [ ] As a Polling Station Officer I can confirm that the physical and digital closure certificate have the same information
- [ ] I see a confirmation dialog when making this action
- [ ] After I confirmed I can't undo anything
| non_priority | sign the digital electoral closure certificate ref v pgmf as the polling station officer i want to be able to sign and validate that the data that i ve uploaded to the web system the digital electoral closure certificate and the one that i have in the physical closure certificate are the same i should have a checkbox or some way that says i ve reviewed this and is the same as the physical electoral closure certificate this is closely related to fill in a web form the results digital electoral closure certificate ref v pgmf capture a picture of the physical electoral closure certificate ref v pgmf as it s the confirmation that both of these certificates have the same information after i ve confirmed it i can no longer modify any of the information so i need to have a confirmation dialog explaining that this action can t be undone acceptance criteria as a polling station officer i can confirm that the physical and digital closure certificate have the same information i see a confirmation dialog when making this action after i confirmed i can t undo anything | 0 |
123,941 | 12,222,515,893 | IssuesEvent | 2020-05-02 13:35:30 | E-Edu/concept | https://api.github.com/repos/E-Edu/concept | opened | Add examples | documentation endpoints | ## Microservice
task
## Priority
high
## Issue
Please add to every parameter and response an example, so we can use Postman to generate a mocking server by the swagger documentation to test our code
If you have any questions please contact @adamane | 1.0 | Add examples - ## Microservice
task
## Priority
high
## Issue
Please add to every parameter and response an example, so we can use Postman to generate a mocking server by the swagger documentation to test our code
If you have any questions please contact @adamane | non_priority | add examples microservice task priority high issue please add to every parameter and response an example so we can use postman to generate a mocking server by the swagger documentation to test our code if you have any questions please contact adamane | 0 |
55,652 | 30,863,349,774 | IssuesEvent | 2023-08-03 06:03:24 | vitessio/vitess | https://api.github.com/repos/vitessio/vitess | opened | Feature Request: improve ApplySchema multi-query performance in `direct` strategy | Type: Feature Request Type: Performance Component: vtctl | ### Feature Description
When `ApplySchema` is supplied with multiple queries, it runs a per-query loop that applies each query onto the shards. With `direct` strategy, we just run `ExecuteFetchAsDBA`.
A common scenario is when you want to bootstrap a new shard/cluster with a known schema. In that scenario, one may need to apply thousands of `CREATE TABLE|VIEW` statements onto the shards. Applying those one-at-a-time takes a significant toll due do network overhead.
We want to improve this performance.,
### Use Case(s)
Populating new large schemas on empty shards. | True | Feature Request: improve ApplySchema multi-query performance in `direct` strategy - ### Feature Description
When `ApplySchema` is supplied with multiple queries, it runs a per-query loop that applies each query onto the shards. With `direct` strategy, we just run `ExecuteFetchAsDBA`.
A common scenario is when you want to bootstrap a new shard/cluster with a known schema. In that scenario, one may need to apply thousands of `CREATE TABLE|VIEW` statements onto the shards. Applying those one-at-a-time takes a significant toll due do network overhead.
We want to improve this performance.,
### Use Case(s)
Populating new large schemas on empty shards. | non_priority | feature request improve applyschema multi query performance in direct strategy feature description when applyschema is supplied with multiple queries it runs a per query loop that applies each query onto the shards with direct strategy we just run executefetchasdba a common scenario is when you want to bootstrap a new shard cluster with a known schema in that scenario one may need to apply thousands of create table view statements onto the shards applying those one at a time takes a significant toll due do network overhead we want to improve this performance use case s populating new large schemas on empty shards | 0 |
256,924 | 19,477,069,765 | IssuesEvent | 2021-12-24 14:51:06 | herbsjs/herbs2mongo | https://api.github.com/repos/herbsjs/herbs2mongo | closed | Wrong documentation | documentation released | the documentation talk about a `database` property that must to be passed to super constructor, but where it is used into herbs2mongo? I read the code but can't find it.
``` js
class ItemRepository extends Repository {
constructor() {
super({
entity: Item,
collection: 'aCollection',
database, // <---- ??????????
ids: ['id'],
mongodb: await connection()
})
}
```
other mistake is that TODO list, the `findAll` is marked as true but it's not implemented | 1.0 | Wrong documentation - the documentation talk about a `database` property that must to be passed to super constructor, but where it is used into herbs2mongo? I read the code but can't find it.
``` js
class ItemRepository extends Repository {
constructor() {
super({
entity: Item,
collection: 'aCollection',
database, // <---- ??????????
ids: ['id'],
mongodb: await connection()
})
}
```
other mistake is that TODO list, the `findAll` is marked as true but it's not implemented | non_priority | wrong documentation the documentation talk about a database property that must to be passed to super constructor but where it is used into i read the code but can t find it js class itemrepository extends repository constructor super entity item collection acollection database ids mongodb await connection other mistake is that todo list the findall is marked as true but it s not implemented | 0 |
33,024 | 15,763,310,069 | IssuesEvent | 2021-03-31 12:04:44 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | TGUI menus laggy | Performance UI | Round ID: https://scrubby.melonmesa.com/round/159099
Reproduction: tgui menus were INCREDIBLY laggy and froze everytime upon opening, the only thing I can think that has caused this (this has been happening every round so far, and other players notice it too) is that possibly a runtime happened, since this same round had cells not charging with power | True | TGUI menus laggy - Round ID: https://scrubby.melonmesa.com/round/159099
Reproduction: tgui menus were INCREDIBLY laggy and froze everytime upon opening, the only thing I can think that has caused this (this has been happening every round so far, and other players notice it too) is that possibly a runtime happened, since this same round had cells not charging with power | non_priority | tgui menus laggy round id reproduction tgui menus were incredibly laggy and froze everytime upon opening the only thing i can think that has caused this this has been happening every round so far and other players notice it too is that possibly a runtime happened since this same round had cells not charging with power | 0 |
275,508 | 23,919,167,626 | IssuesEvent | 2022-09-09 15:12:48 | blockframes/blockframes | https://api.github.com/repos/blockframes/blockframes | closed | e2e test : new auth flow for public, protected and private events | Dev - Test / Quality Assurance July clean up | - [ ] Access to public event by choosing guest role and providing lastname / firstname
- [ ] Access to protected event by choosing guest role and providing lastname / firstname / email (invitationId should be in url)
- [ ] Access to a private event with a regular account
- [ ] A lot more can be imagined
ask @Poppe-A or @phpgeekfr for more details and look at https://www.notion.so/cascade8/All-Test-Scenarios-e4e937051d264f98abe65fe3190bfbe9 | 1.0 | e2e test : new auth flow for public, protected and private events - - [ ] Access to public event by choosing guest role and providing lastname / firstname
- [ ] Access to protected event by choosing guest role and providing lastname / firstname / email (invitationId should be in url)
- [ ] Access to a private event with a regular account
- [ ] A lot more can be imagined
ask @Poppe-A or @phpgeekfr for more details and look at https://www.notion.so/cascade8/All-Test-Scenarios-e4e937051d264f98abe65fe3190bfbe9 | non_priority | test new auth flow for public protected and private events access to public event by choosing guest role and providing lastname firstname access to protected event by choosing guest role and providing lastname firstname email invitationid should be in url access to a private event with a regular account a lot more can be imagined ask poppe a or phpgeekfr for more details and look at | 0 |
147,920 | 19,526,153,749 | IssuesEvent | 2021-12-30 08:12:20 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2016-8645 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2016-8645 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The TCP stack in the Linux kernel before 4.8.10 mishandles skb truncation, which allows local users to cause a denial of service (system crash) via a crafted application that makes sendto system calls, related to net/ipv4/tcp_ipv4.c and net/ipv6/tcp_ipv6.c.
<p>Publish Date: 2016-11-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-8645>CVE-2016-8645</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-8645">https://nvd.nist.gov/vuln/detail/CVE-2016-8645</a></p>
<p>Release Date: 2016-11-28</p>
<p>Fix Resolution: 4.8.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-8645 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-8645 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The TCP stack in the Linux kernel before 4.8.10 mishandles skb truncation, which allows local users to cause a denial of service (system crash) via a crafted application that makes sendto system calls, related to net/ipv4/tcp_ipv4.c and net/ipv6/tcp_ipv6.c.
<p>Publish Date: 2016-11-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-8645>CVE-2016-8645</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-8645">https://nvd.nist.gov/vuln/detail/CVE-2016-8645</a></p>
<p>Release Date: 2016-11-28</p>
<p>Fix Resolution: 4.8.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the tcp stack in the linux kernel before mishandles skb truncation which allows local users to cause a denial of service system crash via a crafted application that makes sendto system calls related to net tcp c and net tcp c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
286,187 | 24,727,674,564 | IssuesEvent | 2022-10-20 15:06:52 | redhat-developer/vscode-openshift-tools | https://api.github.com/repos/redhat-developer/vscode-openshift-tools | opened | Cannot Start Dev action on quarkus component until an OpenShift Project is created | odov3-test | Env. Linux, Fedora 36, OCP: local CRC 2.9.0.
Maybe it is simply not that clear to me, but I could not execute (it does nothing, just wrote out Starting odo dev but nothing else) Start Dev operation on created quarkus component, until a project is created on OpenShift connection. | 1.0 | Cannot Start Dev action on quarkus component until an OpenShift Project is created - Env. Linux, Fedora 36, OCP: local CRC 2.9.0.
Maybe it is simply not that clear to me, but I could not execute (it does nothing, just wrote out Starting odo dev but nothing else) Start Dev operation on created quarkus component, until a project is created on OpenShift connection. | non_priority | cannot start dev action on quarkus component until an openshift project is created env linux fedora ocp local crc maybe it is simply not that clear to me but i could not execute it does nothing just wrote out starting odo dev but nothing else start dev operation on created quarkus component until a project is created on openshift connection | 0 |
209,601 | 16,044,158,879 | IssuesEvent | 2021-04-22 11:42:59 | easydigitaldownloads/easy-digital-downloads | https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads | closed | 3.0 - $order->get_discounts can return an empty array | type-bug workflow-blocker workflow-has-pr workflow-needs-testing | ## Bug Report
### Expected behavior
When I have a proper `$order` from `edd_get_order()`, I expect that `$order->get_discounts()` will return the discounts applied to the order.
### Actual behavior
When I use `$order->get_discounts()` before otherwise getting the order adjustments, no discounts are returned. This is happening with the PR for #8508.
### Steps to reproduce the behavior
1. Check out `issue/8508`.
2. Create an order with a discount applied.
3. View the order receipt/confirmation. The discount does not show.
### Information (if a specific version is affected):
PHP Version: 7.4.1
EDD Version (or branch): release/3.0, but obvious in issue/8508
WordPress Version: 5.7
Any other relevant information:
`get_discounts` [accesses the adjustments property directly](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/release/3.0/includes/orders/classes/class-order.php#L387), but if they have not been retrieved already, no discounts are returned.
I believe the `get_discounts` method needs to be updated to use the `get_adjustments` method the same way the [`get_fees` method](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/release/3.0/includes/orders/classes/class-order.php#L414-L417) does. | 1.0 | 3.0 - $order->get_discounts can return an empty array - ## Bug Report
### Expected behavior
When I have a proper `$order` from `edd_get_order()`, I expect that `$order->get_discounts()` will return the discounts applied to the order.
### Actual behavior
When I use `$order->get_discounts()` before otherwise getting the order adjustments, no discounts are returned. This is happening with the PR for #8508.
### Steps to reproduce the behavior
1. Check out `issue/8508`.
2. Create an order with a discount applied.
3. View the order receipt/confirmation. The discount does not show.
### Information (if a specific version is affected):
PHP Version: 7.4.1
EDD Version (or branch): release/3.0, but obvious in issue/8508
WordPress Version: 5.7
Any other relevant information:
`get_discounts` [accesses the adjustments property directly](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/release/3.0/includes/orders/classes/class-order.php#L387), but if they have not been retrieved already, no discounts are returned.
I believe the `get_discounts` method needs to be updated to use the `get_adjustments` method the same way the [`get_fees` method](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/release/3.0/includes/orders/classes/class-order.php#L414-L417) does. | non_priority | order get discounts can return an empty array bug report expected behavior when i have a proper order from edd get order i expect that order get discounts will return the discounts applied to the order actual behavior when i use order get discounts before otherwise getting the order adjustments no discounts are returned this is happening with the pr for steps to reproduce the behavior check out issue create an order with a discount applied view the order receipt confirmation the discount does not show information if a specific version is affected php version edd version or branch release but obvious in issue wordpress version any other relevant information get discounts but if they have not been retrieved already no discounts are returned i believe the get discounts method needs to be updated to use the get adjustments method the same way the does | 0 |
44,981 | 11,569,487,565 | IssuesEvent | 2020-02-20 17:38:45 | openthread/openthread | https://api.github.com/repos/openthread/openthread | closed | [build] getting error when running make check locally | comp: build continuous-integration | **Describe the bug**
When building as follow locally(linux):
```
./bootstrap
REFERENCE_DEVICE=1 COVERAGE=1 CFLAGS=-m32 CXXFLAGS=-m32 LDFLAGS=-m32 make -f examples/Makefile-posix check
```
I got the such error:
> /tests/unit/test_lowpan.hpp:63:38: error: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type class ot::TestIphcVector’; use assignment instead [-Werror=class-memaccess]
> 63 | memset(this, 0, sizeof(*this));
>
It seems that memset a class object is not allowed. | 1.0 | [build] getting error when running make check locally - **Describe the bug**
When building as follow locally(linux):
```
./bootstrap
REFERENCE_DEVICE=1 COVERAGE=1 CFLAGS=-m32 CXXFLAGS=-m32 LDFLAGS=-m32 make -f examples/Makefile-posix check
```
I got the such error:
> /tests/unit/test_lowpan.hpp:63:38: error: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type class ot::TestIphcVector’; use assignment instead [-Werror=class-memaccess]
> 63 | memset(this, 0, sizeof(*this));
>
It seems that memset a class object is not allowed. | non_priority | getting error when running make check locally describe the bug when building as follow locally linux bootstrap reference device coverage cflags cxxflags ldflags make f examples makefile posix check i got the such error tests unit test lowpan hpp error ‘void memset void int size t ’ clearing an object of non trivial type class ot testiphcvector’ use assignment instead memset this sizeof this it seems that memset a class object is not allowed | 0 |
88,954 | 11,184,006,299 | IssuesEvent | 2019-12-31 15:58:53 | microsoft/pyright | https://api.github.com/repos/microsoft/pyright | closed | assigning member to ModuleType | as designed | **Describe the bug**
I am trying to set a module level variable from within a method. module accesed from `sys.modules` has type `ModuleType`. Settings variable to this raises an error
```
Cannot assign member 'bar' for type 'ModuleType'
Member 'bar' is unknown
```
**To Reproduce**
```
project_root
├── foo
│ └── __init__.py
└── run.py
```
``` python
# Inside foo/__init__.py
import sys
bar = "bar not set yet"
def set_bar():
this_module = sys.modules[__name__] # <-- this_module of type ModuleType
this_module.bar = "new bar value" # <-- pyright error raised in this line
# run.py
import foo
print(foo.bar)
foo.set_bar()
print(foo.bar)
```
**Expected behavior**
Should not raise an error
**VS Code extension or command-line**
extension v 1.1.16
| 1.0 | assigning member to ModuleType - **Describe the bug**
I am trying to set a module level variable from within a method. module accesed from `sys.modules` has type `ModuleType`. Settings variable to this raises an error
```
Cannot assign member 'bar' for type 'ModuleType'
Member 'bar' is unknown
```
**To Reproduce**
```
project_root
├── foo
│ └── __init__.py
└── run.py
```
``` python
# Inside foo/__init__.py
import sys
bar = "bar not set yet"
def set_bar():
this_module = sys.modules[__name__] # <-- this_module of type ModuleType
this_module.bar = "new bar value" # <-- pyright error raised in this line
# run.py
import foo
print(foo.bar)
foo.set_bar()
print(foo.bar)
```
**Expected behavior**
Should not raise an error
**VS Code extension or command-line**
extension v 1.1.16
| non_priority | assigning member to moduletype describe the bug i am trying to set a module level variable from within a method module accesed from sys modules has type moduletype settings variable to this raises an error cannot assign member bar for type moduletype member bar is unknown to reproduce project root ├── foo │ └── init py └── run py python inside foo init py import sys bar bar not set yet def set bar this module sys modules this module of type moduletype this module bar new bar value pyright error raised in this line run py import foo print foo bar foo set bar print foo bar expected behavior should not raise an error vs code extension or command line extension v | 0 |
13,782 | 9,049,380,178 | IssuesEvent | 2019-02-12 04:21:23 | aurmeneta/projektwoche | https://api.github.com/repos/aurmeneta/projektwoche | reopened | No hay botón para cerrar sesión en pantallas pequeñas | enhancement security | En vistas pequeñas, no se puede cerrar la sesión de administrador. | True | No hay botón para cerrar sesión en pantallas pequeñas - En vistas pequeñas, no se puede cerrar la sesión de administrador. | non_priority | no hay botón para cerrar sesión en pantallas pequeñas en vistas pequeñas no se puede cerrar la sesión de administrador | 0 |
6,021 | 6,110,297,149 | IssuesEvent | 2017-06-21 14:46:22 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Running tests with netfx from Visual Studio | area-Infrastructure question | I know that VS isn't officially fully supported in corefx, but is there a way to have the F5/edit, debug continue experience for tests running against netfx? There may be a way to do this with the command line arguments | 1.0 | Running tests with netfx from Visual Studio - I know that VS isn't officially fully supported in corefx, but is there a way to have the F5/edit, debug continue experience for tests running against netfx? There may be a way to do this with the command line arguments | non_priority | running tests with netfx from visual studio i know that vs isn t officially fully supported in corefx but is there a way to have the edit debug continue experience for tests running against netfx there may be a way to do this with the command line arguments | 0 |
51,870 | 10,731,605,340 | IssuesEvent | 2019-10-28 19:56:15 | torilmud/docs | https://api.github.com/repos/torilmud/docs | closed | Auction god view should include item vNum | code enhancement | Just be a nice little add if you could see the object #s going up on auction. | 1.0 | Auction god view should include item vNum - Just be a nice little add if you could see the object #s going up on auction. | non_priority | auction god view should include item vnum just be a nice little add if you could see the object s going up on auction | 0 |
9,839 | 7,009,774,547 | IssuesEvent | 2017-12-19 20:26:48 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | Sites do not load when clicking 'My Sites' | OSS Citizen Performance Sites [Type] Bug | <!-- Thanks for contributing to Calypso! Pick a clear title ("Editor: add spell check") and proceed. -->
A user is reporting that when they click on ‘My Sites’, nothing ever loads, therefore preventing access to any of their sites. It continues to display ‘Loading my sites’. I was able to confirm this after using SU on the user’s account.
Ticket: 816070-zen
#### Steps to reproduce
1. SU into the user's dashboard.
2. Click 'My sites'
#### What I expected
I expected the user's sites to be displayed properly.
#### What happened instead
Nothing loads. It continues to display 'Loading my sites'.
<!-- Optional: share your unique context to help us understand your perspective. You can add context tags such as: #journey #anecdote #narrative #context #empathy #perspective #reallife #dogfooding #20364 esharing #flowsharing #anxiety #anxiety-flow #stresscase #painpoint.
We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable.
If requesting a new feature, explain why you'd like to see it added.
-->
<!--
PLEASE NOTE
- These comments won't show up when you submit the issue.
- Everything is optional, but try to add as many details as possible.
Docs & troubleshooting:
https://github.com/Automattic/wp-calypso/blob/master/.github/CONTRIBUTING.md
https://github.com/Automattic/wp-calypso/blob/master/docs/troubleshooting.md
Helpful tips for screenshots:
https://en.support.wordpress.com/make-a-screenshot/
-->
| True | Sites do not load when clicking 'My Sites' - <!-- Thanks for contributing to Calypso! Pick a clear title ("Editor: add spell check") and proceed. -->
A user is reporting that when they click on ‘My Sites’, nothing ever loads, therefore preventing access to any of their sites. It continues to display ‘Loading my sites’. I was able to confirm this after using SU on the user’s account.
Ticket: 816070-zen
#### Steps to reproduce
1. SU into the user's dashboard.
2. Click 'My sites'
#### What I expected
I expected the user's sites to be displayed properly.
#### What happened instead
Nothing loads. It continues to display 'Loading my sites'.
<!-- Optional: share your unique context to help us understand your perspective. You can add context tags such as: #journey #anecdote #narrative #context #empathy #perspective #reallife #dogfooding #20364 esharing #flowsharing #anxiety #anxiety-flow #stresscase #painpoint.
We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable.
If requesting a new feature, explain why you'd like to see it added.
-->
<!--
PLEASE NOTE
- These comments won't show up when you submit the issue.
- Everything is optional, but try to add as many details as possible.
Docs & troubleshooting:
https://github.com/Automattic/wp-calypso/blob/master/.github/CONTRIBUTING.md
https://github.com/Automattic/wp-calypso/blob/master/docs/troubleshooting.md
Helpful tips for screenshots:
https://en.support.wordpress.com/make-a-screenshot/
-->
| non_priority | sites do not load when clicking my sites a user is reporting that when they click on ‘my sites’ nothing ever loads therefore preventing access to any of their sites it continues to display ‘loading my sites’ i was able to confirm this after using su on the user’s account ticket zen steps to reproduce su into the user s dashboard click my sites what i expected i expected the user s sites to be displayed properly what happened instead nothing loads it continues to display loading my sites optional share your unique context to help us understand your perspective you can add context tags such as journey anecdote narrative context empathy perspective reallife dogfooding esharing flowsharing anxiety anxiety flow stresscase painpoint we d also love to know how you found the bug dogfooding manual testing automated testing or user report if applicable if requesting a new feature explain why you d like to see it added please note these comments won t show up when you submit the issue everything is optional but try to add as many details as possible docs troubleshooting helpful tips for screenshots | 0 |
3,691 | 5,905,945,533 | IssuesEvent | 2017-05-19 14:05:26 | hioa-cs/IncludeOS | https://api.github.com/repos/hioa-cs/IncludeOS | closed | Create system reporting service image | Booting services | .. that can run on any hardware
Prints out CPU capabilities, such as:
- x2APIC support
- AVX and AVX2 support
- Cryptographic extensions (AES, POPCNT)
- SSE3,4.x support
- FMA4
The image will help detect maximum support, and potential problems (eg, low memory)
| 1.0 | Create system reporting service image - .. that can run on any hardware
Prints out CPU capabilities, such as:
- x2APIC support
- AVX and AVX2 support
- Cryptographic extensions (AES, POPCNT)
- SSE3,4.x support
- FMA4
The image will help detect maximum support, and potential problems (eg, low memory)
| non_priority | create system reporting service image that can run on any hardware prints out cpu capabilities such as support avx and support cryptographic extensions aes popcnt x support the image will help detect maximum support and potential problems eg low memory | 0 |
350,803 | 31,932,333,177 | IssuesEvent | 2023-09-19 08:17:02 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_random.test_jax_pareto | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6203010848/job/16842880371"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6203010848/job/16842880371"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_random.test_jax_pareto - | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6203010848/job/16842880371"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6220351767"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6203010848/job/16842880371"><img src=https://img.shields.io/badge/-success-success></a>
| non_priority | fix jax random test jax pareto paddle a href src tensorflow a href src numpy a href src jax a href src torch a href src | 0 |
305,282 | 26,375,723,059 | IssuesEvent | 2023-01-12 02:08:47 | longhorn/longhorn | https://api.github.com/repos/longhorn/longhorn | opened | [TEST] Automate node reboot test cases | kind/test | ## What's the test to develop? Please describe
We need to automate the node reboot test cases as listed below
## Describe the items of the test development (DoD, definition of done) you'd like
> Please use a task list for items on a separate line with a clickable checkbox https://docs.github.com/en/issues/tracking-your-work-with-issues/about-task-lists
Node reboot while the below operation is in progress
- [ ] Replica is rebuilding.
- [ ] Volume migration is in progress.
- [ ] Backup restoring is in progress.
- [ ] Writing of data is in progress.
- [ ] Replica deletion is in progress.
- [ ] Online volume is in progress.
## Additional context
| 1.0 | [TEST] Automate node reboot test cases - ## What's the test to develop? Please describe
We need to automate the node reboot test cases as listed below
## Describe the items of the test development (DoD, definition of done) you'd like
> Please use a task list for items on a separate line with a clickable checkbox https://docs.github.com/en/issues/tracking-your-work-with-issues/about-task-lists
Node reboot while the below operation is in progress
- [ ] Replica is rebuilding.
- [ ] Volume migration is in progress.
- [ ] Backup restoring is in progress.
- [ ] Writing of data is in progress.
- [ ] Replica deletion is in progress.
- [ ] Online volume is in progress.
## Additional context
| non_priority | automate node reboot test cases what s the test to develop please describe we need to automate the node reboot test cases as listed below describe the items of the test development dod definition of done you d like please use a task list for items on a separate line with a clickable checkbox node reboot while the below operation is in progress replica is rebuilding volume migration is in progress backup restoring is in progress writing of data is in progress replica deletion is in progress online volume is in progress additional context | 0 |
91,935 | 18,755,254,177 | IssuesEvent | 2021-11-05 09:54:35 | arduino/arduino-cli | https://api.github.com/repos/arduino/arduino-cli | opened | Use Cobra's PreRun function to parse the arguments | type: enhancement topic: code topic: CLI | We already use the PreRun [here](https://github.com/arduino/arduino-cli/blob/64cb753c944e31664aace944df788bda5a8986a6/cli/upload/upload.go#L52)
We could leverage it to consume the args given to a command before running the actual run function
https://github.com/spf13/cobra/blob/master/user_guide.md#prerun-and-postrun-hooks | 1.0 | Use Cobra's PreRun function to parse the arguments - We already use the PreRun [here](https://github.com/arduino/arduino-cli/blob/64cb753c944e31664aace944df788bda5a8986a6/cli/upload/upload.go#L52)
We could leverage it to consume the args given to a command before running the actual run function
https://github.com/spf13/cobra/blob/master/user_guide.md#prerun-and-postrun-hooks | non_priority | use cobra s prerun function to parse the arguments we already use the prerun we could leverage it to consume the args given to a command before running the actual run function | 0 |
18,832 | 4,314,083,329 | IssuesEvent | 2016-07-22 13:18:24 | arekbulski/Cameleonica | https://api.github.com/repos/arekbulski/Cameleonica | closed | Make a structure for PDF viewers | Documentation | When viewing PDF version there are not many labels to jump to.

This problem is already posted on AskUbuntu:
https://askubuntu.com/questions/606093/libreoffice-writer-adding-bookmarks
| 1.0 | Make a structure for PDF viewers - When viewing PDF version there are not many labels to jump to.

This problem is already posted on AskUbuntu:
https://askubuntu.com/questions/606093/libreoffice-writer-adding-bookmarks
| non_priority | make a structure for pdf viewers when viewing pdf version there are not many labels to jump to this problem is already posted on askubuntu | 0 |
72,849 | 19,521,075,275 | IssuesEvent | 2021-12-29 18:34:36 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | opened | [Task]: Add keyboard focus state to canvas size selector | UI Builders Pod Task | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks

| 1.0 | [Task]: Add keyboard focus state to canvas size selector - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks

| non_priority | add keyboard focus state to canvas size selector is there an existing issue for this i have searched the existing issues subtasks | 0 |
100,832 | 21,525,787,007 | IssuesEvent | 2022-04-28 18:17:50 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [Perf] Changes at 4/3/2022 6:13:34 PM | area-CodeGen-coreclr refs/heads/main RunKind=micro Windows 10.0.18362 Regression CoreClr x86 | ### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in PerfLabTests.CastingPerf2.CastingPerf
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IFooObjIsIFoo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/PerfLabTests.CastingPerf2.CastingPerf.IFooObjIsIFoo.html>) | 533.95 μs | 585.14 μs | 1.10 | 0.04 | False | | |
[FooObjIsDescendant - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/PerfLabTests.CastingPerf2.CastingPerf.FooObjIsDescendant.html>) | 408.03 μs | 487.88 μs | 1.20 | 0.18 | False | | |


[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/PerfLabTests.CastingPerf2.CastingPerf.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'PerfLabTests.CastingPerf2.CastingPerf*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### PerfLabTests.CastingPerf2.CastingPerf.IFooObjIsIFoo
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 585.142199074074 > 560.6315528445513.
IsChangePoint: Marked as a change because one of 3/10/2022 1:12:51 AM, 3/22/2022 2:38:09 PM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -26.073826926551938 (T) = (0 -585079.1995517344) / Math.Sqrt((122660132.96965332 / (49)) + (2379427.143725326 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.08212072884065917 = (540678.3032227511 - 585079.1995517344) / 540678.3032227511 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### PerfLabTests.CastingPerf2.CastingPerf.FooObjIsDescendant
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 487.88003246753243 > 428.6117251602564.
IsChangePoint: Marked as a change because one of 2/25/2022 1:44:54 AM, 3/4/2022 7:00:20 AM, 3/7/2022 10:45:01 PM, 3/16/2022 5:02:20 PM, 3/21/2022 1:28:14 PM, 3/31/2022 5:34:47 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -17.992324298006597 (T) = (0 -477791.5116915254) / Math.Sqrt((57977535.77841977 / (35)) + (250451561.61398742 / (20))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (35) + (20) - 2, .025) and -0.16522805280935102 = (410041.2022690114 - 477791.5116915254) / 410041.2022690114 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in System.Collections.IndexerSet<Int32>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[ConcurrentDictionary - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/System.Collections.IndexerSet(Int32).ConcurrentDictionary(Size%3a%20512).html>) | 20.22 μs | 24.85 μs | 1.23 | 0.16 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/System.Collections.IndexerSet(Int32).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.IndexerSet<Int32>*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### System.Collections.IndexerSet<Int32>.ConcurrentDictionary(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 24.84945149172474 > 21.34707510538262.
IsChangePoint: Marked as a change because one of 3/7/2022 10:45:01 PM, 3/9/2022 10:22:25 AM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -10.91617552412988 (T) = (0 -24360.34486246879) / Math.Sqrt((268967.87390529225 / (49)) + (712236.7726556616 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.18753689218135378 = (20513.33733112236 - 24360.34486246879) / 20513.33733112236 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in System.Collections.ContainsKeyFalse<String, String>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IDictionary - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/System.Collections.ContainsKeyFalse(String%2c%20String).IDictionary(Size%3a%20512).html>) | 17.66 μs | 20.17 μs | 1.14 | 0.01 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/System.Collections.ContainsKeyFalse(String,%20String).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.ContainsKeyFalse<String, String>*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### System.Collections.ContainsKeyFalse<String, String>.IDictionary(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 20.166918817204298 > 18.506320741572186.
IsChangePoint: Marked as a change because one of 3/17/2022 11:24:40 PM, 3/31/2022 5:34:47 PM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -36.0126424900174 (T) = (0 -20143.384864486816) / Math.Sqrt((164629.9030833987 / (49)) + (5577.151301824765 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.13261720640768843 = (17784.812689165658 - 20143.384864486816) / 17784.812689165658 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| 1.0 | [Perf] Changes at 4/3/2022 6:13:34 PM - ### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in PerfLabTests.CastingPerf2.CastingPerf
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IFooObjIsIFoo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/PerfLabTests.CastingPerf2.CastingPerf.IFooObjIsIFoo.html>) | 533.95 μs | 585.14 μs | 1.10 | 0.04 | False | | |
[FooObjIsDescendant - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/PerfLabTests.CastingPerf2.CastingPerf.FooObjIsDescendant.html>) | 408.03 μs | 487.88 μs | 1.20 | 0.18 | False | | |


[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/PerfLabTests.CastingPerf2.CastingPerf.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'PerfLabTests.CastingPerf2.CastingPerf*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### PerfLabTests.CastingPerf2.CastingPerf.IFooObjIsIFoo
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 585.142199074074 > 560.6315528445513.
IsChangePoint: Marked as a change because one of 3/10/2022 1:12:51 AM, 3/22/2022 2:38:09 PM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -26.073826926551938 (T) = (0 -585079.1995517344) / Math.Sqrt((122660132.96965332 / (49)) + (2379427.143725326 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.08212072884065917 = (540678.3032227511 - 585079.1995517344) / 540678.3032227511 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### PerfLabTests.CastingPerf2.CastingPerf.FooObjIsDescendant
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 487.88003246753243 > 428.6117251602564.
IsChangePoint: Marked as a change because one of 2/25/2022 1:44:54 AM, 3/4/2022 7:00:20 AM, 3/7/2022 10:45:01 PM, 3/16/2022 5:02:20 PM, 3/21/2022 1:28:14 PM, 3/31/2022 5:34:47 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -17.992324298006597 (T) = (0 -477791.5116915254) / Math.Sqrt((57977535.77841977 / (35)) + (250451561.61398742 / (20))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (35) + (20) - 2, .025) and -0.16522805280935102 = (410041.2022690114 - 477791.5116915254) / 410041.2022690114 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in System.Collections.IndexerSet<Int32>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[ConcurrentDictionary - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/System.Collections.IndexerSet(Int32).ConcurrentDictionary(Size%3a%20512).html>) | 20.22 μs | 24.85 μs | 1.23 | 0.16 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/System.Collections.IndexerSet(Int32).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.IndexerSet<Int32>*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### System.Collections.IndexerSet<Int32>.ConcurrentDictionary(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 24.84945149172474 > 21.34707510538262.
IsChangePoint: Marked as a change because one of 3/7/2022 10:45:01 PM, 3/9/2022 10:22:25 AM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -10.91617552412988 (T) = (0 -24360.34486246879) / Math.Sqrt((268967.87390529225 / (49)) + (712236.7726556616 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.18753689218135378 = (20513.33733112236 - 24360.34486246879) / 20513.33733112236 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | x86
-- | --
OS | Windows 10.0.18362
Baseline | [a0f7927c0ce4cfa8d1c832e70461b0145389a8be](https://github.com/dotnet/runtime/commit/a0f7927c0ce4cfa8d1c832e70461b0145389a8be)
Compare | [0b4af007f758b7f265a54565251c633b632cc999](https://github.com/dotnet/runtime/commit/0b4af007f758b7f265a54565251c633b632cc999)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a0f7927c0ce4cfa8d1c832e70461b0145389a8be...0b4af007f758b7f265a54565251c633b632cc999)
### Regressions in System.Collections.ContainsKeyFalse<String, String>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IDictionary - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x86_Windows 10.0.18362/System.Collections.ContainsKeyFalse(String%2c%20String).IDictionary(Size%3a%20512).html>) | 17.66 μs | 20.17 μs | 1.14 | 0.01 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_05_2022/refs/heads/main_x86_Windows%2010.0.18362_Regression/System.Collections.ContainsKeyFalse(String,%20String).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.ContainsKeyFalse<String, String>*'
```
<details>
### Payloads
[Baseline](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-64196150-3173-4003-af32-a91d9d75fdffef784626c854992b6/98258d96-0c24-488d-b3be-36a1f7d35290.zip?sv=2019-07-07&se=2022-05-03T10%3A02%3A06Z&sr=c&sp=rl&sig=c7RJXwUZrJa4BH2L5vleagzQF8q6SyNfq8OF5UYIDog%3D>)
[Compare](<https://helixdi8s23ayyeko0k025g8.blob.core.windows.net/helix-job-daf4b595-b65c-4db0-8f82-9fef1c1fe7ae435002e6db4437da4/ed927be9-61b3-4dbe-a77c-26f1485041d3.zip?sv=2019-07-07&se=2022-05-04T00%3A20%3A10Z&sr=c&sp=rl&sig=wtTZ0sRjC5LfuEnZ2V%2FS%2Fey7DqTXq0DRs40%2Fafjq35c%3D>)
### Histogram
#### System.Collections.ContainsKeyFalse<String, String>.IDictionary(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 20.166918817204298 > 18.506320741572186.
IsChangePoint: Marked as a change because one of 3/17/2022 11:24:40 PM, 3/31/2022 5:34:47 PM, 4/3/2022 1:52:36 PM, 4/5/2022 2:02:34 AM falls between 3/26/2022 9:07:40 PM and 4/5/2022 2:02:34 AM.
IsRegressionStdDev: Marked as regression because -36.0126424900174 (T) = (0 -20143.384864486816) / Math.Sqrt((164629.9030833987 / (49)) + (5577.151301824765 / (6))) is less than -2.005745995316835 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (49) + (6) - 2, .025) and -0.13261720640768843 = (17784.812689165658 - 20143.384864486816) / 17784.812689165658 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| non_priority | changes at pm run information architecture os windows baseline compare diff regressions in perflabtests castingperf benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl μs μs false μs μs false repro cmd git clone py performance scripts benchmarks ci py f filter perflabtests castingperf payloads histogram perflabtests castingperf ifooobjisifoo log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am pm pm am falls between pm and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so perflabtests castingperf fooobjisdescendant log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am am pm pm pm pm am falls between pm and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs run information architecture os windows baseline compare diff regressions in system collections indexerset lt gt benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl μs μs false repro cmd git clone py performance scripts benchmarks ci py f filter system collections indexerset lt gt payloads histogram system collections indexerset lt gt concurrentdictionary size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm am pm am falls between pm and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs run information architecture os windows baseline compare diff regressions in system collections containskeyfalse lt string string gt benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl μs μs false repro cmd git clone py performance scripts benchmarks ci py f filter system collections containskeyfalse lt string string gt payloads histogram system collections containskeyfalse lt string string gt idictionary size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm pm am falls between pm and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs | 0 |
239,157 | 26,205,986,214 | IssuesEvent | 2023-01-03 22:42:01 | BranislavBeno/GitLab-Issue-Importer | https://api.github.com/repos/BranislavBeno/GitLab-Issue-Importer | closed | snakeyaml-1.33.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.33.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.33/2cd0a87ff7df953f810c344bdf2fe3340b954c69/snakeyaml-1.33.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (snakeyaml version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.33/2cd0a87ff7df953f810c344bdf2fe3340b954c69/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | snakeyaml-1.33.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.33.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.33/2cd0a87ff7df953f810c344bdf2fe3340b954c69/snakeyaml-1.33.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (snakeyaml version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.33/2cd0a87ff7df953f810c344bdf2fe3340b954c69/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_priority | snakeyaml jar vulnerabilities highest severity is autoclosed vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org yaml snakeyaml snakeyaml jar vulnerabilities cve severity cvss dependency type fixed in snakeyaml version remediation available high snakeyaml jar direct n a details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library found in base branch main vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
99,833 | 16,459,693,649 | IssuesEvent | 2021-05-21 17:00:12 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | web api with microsoft identity error 401 | Author: Migration Bot :robot: Needs: Author Feedback Status: No Recent Activity area-security | _This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/t/web-api-with-microsoft-identity-error-40/1422358)._
---
web api with microsoft identity error 401
Curl
curl -X GET "https://localhost:44395/WeatherForecast" -H "accept: text/plain"
Request URL
https://localhost:44395/WeatherForecast
Server response
Code Details
401
Undocumented
Error:
Response headers
date: Wed12 May 2021 21:49:15 GMT
server: Microsoft-IIS/10.0
www-authenticate: Bearer
x-powered-by: ASP.NET
---
### Original Comments
#### Feedback Bot on 5/12/2021, 07:56 PM:
<p>We have directed your feedback to the appropriate engineering team for further evaluation. The team will review the feedback and notify you about the next steps.</p>
#### Feedback Bot on 5/13/2021, 00:24 AM:
<p>Thank you for sharing your feedback! Our teams prioritize action on product issues with broad customer impact. See details at: <a target='_blank' href="https://docs.microsoft.com/en-us/visualstudio/ide/report-a-problem?view=vs-2019#faq">https://docs.microsoft.com/en-us/visualstudio/ide/report-a-problem?view=vs-2019#faq</a>. In case you need answers to common questions or need assisted support, be sure to use <a target='_blank' href="https://visualstudio.microsoft.com/vs/support/">https://visualstudio.microsoft.com/vs/support/</a>. We’ll keep you posted on any updates to this feedback.</p>
---
### Original Solutions
(no solutions) | True | web api with microsoft identity error 401 - _This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/t/web-api-with-microsoft-identity-error-40/1422358)._
---
web api with microsoft identity error 401
Curl
curl -X GET "https://localhost:44395/WeatherForecast" -H "accept: text/plain"
Request URL
https://localhost:44395/WeatherForecast
Server response
Code Details
401
Undocumented
Error:
Response headers
date: Wed12 May 2021 21:49:15 GMT
server: Microsoft-IIS/10.0
www-authenticate: Bearer
x-powered-by: ASP.NET
---
### Original Comments
#### Feedback Bot on 5/12/2021, 07:56 PM:
<p>We have directed your feedback to the appropriate engineering team for further evaluation. The team will review the feedback and notify you about the next steps.</p>
#### Feedback Bot on 5/13/2021, 00:24 AM:
<p>Thank you for sharing your feedback! Our teams prioritize action on product issues with broad customer impact. See details at: <a target='_blank' href="https://docs.microsoft.com/en-us/visualstudio/ide/report-a-problem?view=vs-2019#faq">https://docs.microsoft.com/en-us/visualstudio/ide/report-a-problem?view=vs-2019#faq</a>. In case you need answers to common questions or need assisted support, be sure to use <a target='_blank' href="https://visualstudio.microsoft.com/vs/support/">https://visualstudio.microsoft.com/vs/support/</a>. We’ll keep you posted on any updates to this feedback.</p>
---
### Original Solutions
(no solutions) | non_priority | web api with microsoft identity error this issue has been moved from web api with microsoft identity error curl curl x get h accept text plain request url server response code details undocumented error response headers date may gmt server microsoft iis www authenticate bearer x powered by asp net original comments feedback bot on pm we have directed your feedback to the appropriate engineering team for further evaluation the team will review the feedback and notify you about the next steps feedback bot on am thank you for sharing your feedback our teams prioritize action on product issues with broad customer impact see details at original solutions no solutions | 0 |
19,025 | 3,737,603,466 | IssuesEvent | 2016-03-08 19:51:10 | EasyRPG/Player | https://api.github.com/repos/EasyRPG/Player | closed | Interpreter hangs when running "continue processing events" command after "move event" command execution on parallel interpreter | Events Hang Interpreter Patch available Testcase available | How to reproduce:
1. Create a parallel event with a move event command.
2. On the same map, create an automatic event with a continue processing events command.
3. This command must be set after some behavior specific commands, like wait command or move event. Otherwise, after e.g. a show message will work fine (?).
Here is a [test case](https://fran.cr/owncloud/index.php/s/KHwW2DiMk4IZoz2/download) to verify this weird race condition.
Verified affected games:
- Desolate City (initial cutscene) | 1.0 | Interpreter hangs when running "continue processing events" command after "move event" command execution on parallel interpreter - How to reproduce:
1. Create a parallel event with a move event command.
2. On the same map, create an automatic event with a continue processing events command.
3. This command must be set after some behavior specific commands, like wait command or move event. Otherwise, after e.g. a show message will work fine (?).
Here is a [test case](https://fran.cr/owncloud/index.php/s/KHwW2DiMk4IZoz2/download) to verify this weird race condition.
Verified affected games:
- Desolate City (initial cutscene) | non_priority | interpreter hangs when running continue processing events command after move event command execution on parallel interpreter how to reproduce create a parallel event with a move event command on the same map create an automatic event with a continue processing events command this command must be set after some behavior specific commands like wait command or move event otherwise after e g a show message will work fine here is a to verify this weird race condition verified affected games desolate city initial cutscene | 0 |
72,633 | 19,393,487,492 | IssuesEvent | 2021-12-18 00:09:16 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | closed | running examples in CI got a lot slower since using the new renderer | A-Build-System C-Performance | ## Bevy version
Latest from main `add1818a05ad4b8d4672197afd477ab29f8446b5`
## Operating system & version
Ubuntu from GitHub actions
## What you did
Run CI
## What you expected to happen
Examples run in an acceptable time (1 mins to build, 8 mins to run)
## What actually happened
As noted by @cart in https://github.com/bevyengine/bevy/pull/3312#issuecomment-993133114, it takes more time to run examples: 1 mins to build, 33 mins to run
## Additional information
Before the merge: https://github.com/bevyengine/bevy/runs/4499607967?check_suite_focus=true
After the merge: https://github.com/bevyengine/bevy/runs/4532054231?check_suite_focus=true
`run-examples` CI should be made faster, or it should run less examples. Ideally it's run time should be around 10 minutes total
| 1.0 | running examples in CI got a lot slower since using the new renderer - ## Bevy version
Latest from main `add1818a05ad4b8d4672197afd477ab29f8446b5`
## Operating system & version
Ubuntu from GitHub actions
## What you did
Run CI
## What you expected to happen
Examples run in an acceptable time (1 mins to build, 8 mins to run)
## What actually happened
As noted by @cart in https://github.com/bevyengine/bevy/pull/3312#issuecomment-993133114, it takes more time to run examples: 1 mins to build, 33 mins to run
## Additional information
Before the merge: https://github.com/bevyengine/bevy/runs/4499607967?check_suite_focus=true
After the merge: https://github.com/bevyengine/bevy/runs/4532054231?check_suite_focus=true
`run-examples` CI should be made faster, or it should run less examples. Ideally it's run time should be around 10 minutes total
| non_priority | running examples in ci got a lot slower since using the new renderer bevy version latest from main operating system version ubuntu from github actions what you did run ci what you expected to happen examples run in an acceptable time mins to build mins to run what actually happened as noted by cart in it takes more time to run examples mins to build mins to run additional information before the merge after the merge run examples ci should be made faster or it should run less examples ideally it s run time should be around minutes total | 0 |
275,860 | 20,959,851,452 | IssuesEvent | 2022-03-27 16:31:37 | garthmortensen/code_club | https://api.github.com/repos/garthmortensen/code_club | opened | add initials to ascii art header | documentation | cool dude, you dont have your initials in the `cool code` ascii art header. FIX THIS IMMEDIATELY1!! | 1.0 | add initials to ascii art header - cool dude, you dont have your initials in the `cool code` ascii art header. FIX THIS IMMEDIATELY1!! | non_priority | add initials to ascii art header cool dude you dont have your initials in the cool code ascii art header fix this | 0 |
78,562 | 10,073,151,501 | IssuesEvent | 2019-07-24 08:59:12 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Blocks tutoiral misordered | [Type] Documentation | **Describe the bug**
introducing-attributes-and-editable-fields.md and block-controls-toolbar-and-sidebar.md are not in the correct order on the tutorial website
**To reproduce**
Steps to reproduce the behavior:
1. Go to https://developer.wordpress.org/block-editor/tutorials/block-tutorial/
2. See in the sidebar that the two aforementioned tutorial links are not in the correct order
**Expected behavior**
introducing-attributes-and-editable-fields.md should be before block-controls-toolbar-and-sidebar.md | 1.0 | Blocks tutoiral misordered - **Describe the bug**
introducing-attributes-and-editable-fields.md and block-controls-toolbar-and-sidebar.md are not in the correct order on the tutorial website
**To reproduce**
Steps to reproduce the behavior:
1. Go to https://developer.wordpress.org/block-editor/tutorials/block-tutorial/
2. See in the sidebar that the two aforementioned tutorial links are not in the correct order
**Expected behavior**
introducing-attributes-and-editable-fields.md should be before block-controls-toolbar-and-sidebar.md | non_priority | blocks tutoiral misordered describe the bug introducing attributes and editable fields md and block controls toolbar and sidebar md are not in the correct order on the tutorial website to reproduce steps to reproduce the behavior go to see in the sidebar that the two aforementioned tutorial links are not in the correct order expected behavior introducing attributes and editable fields md should be before block controls toolbar and sidebar md | 0 |
24,241 | 5,039,941,017 | IssuesEvent | 2016-12-19 01:37:44 | okTurtles/group-income-simple | https://api.github.com/repos/okTurtles/group-income-simple | opened | Set up GPG signing + verification for all commits + tags | Kind:Documentation Kind:Process Level:Advanced Note:Security Note:Tooling | ### Problem
Baddies.
### Solution
GPG. But make it simple. | 1.0 | Set up GPG signing + verification for all commits + tags - ### Problem
Baddies.
### Solution
GPG. But make it simple. | non_priority | set up gpg signing verification for all commits tags problem baddies solution gpg but make it simple | 0 |
31,878 | 6,652,628,487 | IssuesEvent | 2017-09-29 02:51:46 | aleph7/Upsurge | https://api.github.com/repos/aleph7/Upsurge | closed | ComplexArray Infix operator *= causing crash | t1: defect | I am attempting to run a project with a real to complex forward FFT operation, and it has a 100% repro EXC_BAD_ACCESS crash on the scaling at the end of the the function.
```swift
let count = 64
let frequency = 4.0
let step = 2.0 * Double.pi / Double(count)
let x = ValueArray<Double>((0..<count).map({ step * Double($0) * frequency }))
let fft = FFTDouble(inputLength: x.count)
let complex = fft.forward(sin(x)) // EXC_BAD_ACCESS here
for value in complex {
print(value)
}
```
The crash is found on line 75 of `FFT.swift`, on the `results *= scale`. As of now this means that real to complex FFTs are broken. | 1.0 | ComplexArray Infix operator *= causing crash - I am attempting to run a project with a real to complex forward FFT operation, and it has a 100% repro EXC_BAD_ACCESS crash on the scaling at the end of the the function.
```swift
let count = 64
let frequency = 4.0
let step = 2.0 * Double.pi / Double(count)
let x = ValueArray<Double>((0..<count).map({ step * Double($0) * frequency }))
let fft = FFTDouble(inputLength: x.count)
let complex = fft.forward(sin(x)) // EXC_BAD_ACCESS here
for value in complex {
print(value)
}
```
The crash is found on line 75 of `FFT.swift`, on the `results *= scale`. As of now this means that real to complex FFTs are broken. | non_priority | complexarray infix operator causing crash i am attempting to run a project with a real to complex forward fft operation and it has a repro exc bad access crash on the scaling at the end of the the function swift let count let frequency let step double pi double count let x valuearray count map step double frequency let fft fftdouble inputlength x count let complex fft forward sin x exc bad access here for value in complex print value the crash is found on line of fft swift on the results scale as of now this means that real to complex ffts are broken | 0 |
76,832 | 21,577,794,523 | IssuesEvent | 2022-05-02 15:24:32 | root-project/root | https://api.github.com/repos/root-project/root | closed | Failure to build master branch due to xrootd SHA256 checksum mismatch | bug affects:master in:Build System | - [x] Checked for duplicates
<!--
Please search in
* [GitHub](https://github.com/root-project/root/issues?q=is%3Aissue)
* AND [Jira](https://sft.its.cern.ch/jira/issues/?jql=project %3D ROOT)
for existing reports of your issue.
If you find one, you are very welcome to add to the existing report, for instance "issue still exists in today's master".
-->
### Describe the bug
<!--
A clear and concise description of what the wrong behavior is.
-->
I currently cannot build the master branch due to the fact that the hardcoded SHA256 checksum in /builtins/xrootd/CMakeLists.txt is not the same checksum as the one obtained when downloading the file. There is a mismatch. There is a check that makes the build fail in this case.
### Expected behavior
<!--
A clear and concise description of what you expected to happen.
-->
The checksum in /builtins/xrootd/CMakeLists.txt is the same as the one obtained from the actually downloaded file
### To Reproduce
<!--
Steps to reproduce the behavior:
1. Your code that triggers the issue: at least a part; ideally something we can run ourselves.
2. Don't forget to attach the required input files!
3. How to run your code and / or build it, e.g. `root myMacro.C`, ...
-->
```
git clone https://github.com/root-project/root.git
mkdir build
cd build
cmake ../root/ -Droofit_multiprocess=ON -Dminuit2=ON -Dtesting=ON -DCMAKE_IGNORE_PATH="/home/zef/miniconda3/bin;/home/zef/miniconda3/include;/home/zef/miniconda3/lib" -Dccache=ON -DPYTHON_EXECUTABLE="/usr/bin/python3" -Dbuiltin_fftw3=On
```
Note that the checksum obtained by `sha256sum xrootd-5.4.2.tar.gz` is: 2b394270c55cb3d14f3c44b692311db99d2d25278882b513d4d564b68a58ed4a
The checksum hardcoded in the CMAKE file is: 84e8a9a2bcad116df479f94e985c287dd99fbac0613d4fbb61f4ccc0cef81fa3
### Setup
<!--
1. ROOT version
2. Operating system
3. How you obtained ROOT, such as `dnf install` / binary download / you built it yourself.
-->
ROOT version: master branch as of now
Operating system: Pop!OS
Obtained root through git clone of master branch
### Additional context
<!--
Add any other context about the problem here.
-->
| 1.0 | Failure to build master branch due to xrootd SHA256 checksum mismatch - - [x] Checked for duplicates
<!--
Please search in
* [GitHub](https://github.com/root-project/root/issues?q=is%3Aissue)
* AND [Jira](https://sft.its.cern.ch/jira/issues/?jql=project %3D ROOT)
for existing reports of your issue.
If you find one, you are very welcome to add to the existing report, for instance "issue still exists in today's master".
-->
### Describe the bug
<!--
A clear and concise description of what the wrong behavior is.
-->
I currently cannot build the master branch due to the fact that the hardcoded SHA256 checksum in /builtins/xrootd/CMakeLists.txt is not the same checksum as the one obtained when downloading the file. There is a mismatch. There is a check that makes the build fail in this case.
### Expected behavior
<!--
A clear and concise description of what you expected to happen.
-->
The checksum in /builtins/xrootd/CMakeLists.txt is the same as the one obtained from the actually downloaded file
### To Reproduce
<!--
Steps to reproduce the behavior:
1. Your code that triggers the issue: at least a part; ideally something we can run ourselves.
2. Don't forget to attach the required input files!
3. How to run your code and / or build it, e.g. `root myMacro.C`, ...
-->
```
git clone https://github.com/root-project/root.git
mkdir build
cd build
cmake ../root/ -Droofit_multiprocess=ON -Dminuit2=ON -Dtesting=ON -DCMAKE_IGNORE_PATH="/home/zef/miniconda3/bin;/home/zef/miniconda3/include;/home/zef/miniconda3/lib" -Dccache=ON -DPYTHON_EXECUTABLE="/usr/bin/python3" -Dbuiltin_fftw3=On
```
Note that the checksum obtained by `sha256sum xrootd-5.4.2.tar.gz` is: 2b394270c55cb3d14f3c44b692311db99d2d25278882b513d4d564b68a58ed4a
The checksum hardcoded in the CMAKE file is: 84e8a9a2bcad116df479f94e985c287dd99fbac0613d4fbb61f4ccc0cef81fa3
### Setup
<!--
1. ROOT version
2. Operating system
3. How you obtained ROOT, such as `dnf install` / binary download / you built it yourself.
-->
ROOT version: master branch as of now
Operating system: Pop!OS
Obtained root through git clone of master branch
### Additional context
<!--
Add any other context about the problem here.
-->
| non_priority | failure to build master branch due to xrootd checksum mismatch checked for duplicates please search in and root for existing reports of your issue if you find one you are very welcome to add to the existing report for instance issue still exists in today s master describe the bug a clear and concise description of what the wrong behavior is i currently cannot build the master branch due to the fact that the hardcoded checksum in builtins xrootd cmakelists txt is not the same checksum as the one obtained when downloading the file there is a mismatch there is a check that makes the build fail in this case expected behavior a clear and concise description of what you expected to happen the checksum in builtins xrootd cmakelists txt is the same as the one obtained from the actually downloaded file to reproduce steps to reproduce the behavior your code that triggers the issue at least a part ideally something we can run ourselves don t forget to attach the required input files how to run your code and or build it e g root mymacro c git clone mkdir build cd build cmake root droofit multiprocess on on dtesting on dcmake ignore path home zef bin home zef include home zef lib dccache on dpython executable usr bin dbuiltin on note that the checksum obtained by xrootd tar gz is the checksum hardcoded in the cmake file is setup root version operating system how you obtained root such as dnf install binary download you built it yourself root version master branch as of now operating system pop os obtained root through git clone of master branch additional context add any other context about the problem here | 0 |
4,797 | 2,754,649,310 | IssuesEvent | 2015-04-25 21:42:25 | hardscrabble/hardscrabble.github.io | https://api.github.com/repos/hardscrabble/hardscrabble.github.io | closed | Fix the busted sidenotes | design improvement | It was fun to implement them but they're not great, and they're definitely not responsivie
Maybe switch to http://acdlite.github.io/jquery.sidenotes/ which does the same but better (include a slick responsiveness)
My only hesitation is that *that post* (http://www.hardscrabble.net/2013/sidenotes/) is all about my implementation of sidenotes... Maybe try to include that implementation *only on that post* ?? Is that bonkers? | 1.0 | Fix the busted sidenotes - It was fun to implement them but they're not great, and they're definitely not responsivie
Maybe switch to http://acdlite.github.io/jquery.sidenotes/ which does the same but better (include a slick responsiveness)
My only hesitation is that *that post* (http://www.hardscrabble.net/2013/sidenotes/) is all about my implementation of sidenotes... Maybe try to include that implementation *only on that post* ?? Is that bonkers? | non_priority | fix the busted sidenotes it was fun to implement them but they re not great and they re definitely not responsivie maybe switch to which does the same but better include a slick responsiveness my only hesitation is that that post is all about my implementation of sidenotes maybe try to include that implementation only on that post is that bonkers | 0 |
18,507 | 5,640,170,222 | IssuesEvent | 2017-04-06 15:48:42 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Media Manager not outputing any error when trying to upload a zip file that contains php files | No Code Attached Yet | As far as I understand there is a new security feature that scans the content of a zip file uploaded through the media manager and forbids the upload if that zip file contains any php files.
In my case this is pretty stupid as I want people to upload zip files containing php code, but anyway. That's not the point of this issue here.
When uploading such an "incorrect" file I would expect to at least get an error message or at a warning. Instead the page just refreshes and no message is displayed.
#### Steps to reproduce the issue
Go to media manager. (make sure that you have set the zip file extension in the allowed extension types) Try to upload a zip file containing php files.
#### Expected result
The file should upload.
#### Actual result
The file doesn't upload and not error message is shown. The file list just refreshes.
#### System information (as much as possible)
Tested on several 3.5.1 installations
#### Additional comments
I tried to upload the test zip files here but GitHub says "unfortunately, we don't support that file type", which is also pretty confusing as ZIP is listed as a supported file type, but it's way better than joomla's omission of an error whatsoever.
| 1.0 | Media Manager not outputing any error when trying to upload a zip file that contains php files - As far as I understand there is a new security feature that scans the content of a zip file uploaded through the media manager and forbids the upload if that zip file contains any php files.
In my case this is pretty stupid as I want people to upload zip files containing php code, but anyway. That's not the point of this issue here.
When uploading such an "incorrect" file I would expect to at least get an error message or at a warning. Instead the page just refreshes and no message is displayed.
#### Steps to reproduce the issue
Go to media manager. (make sure that you have set the zip file extension in the allowed extension types) Try to upload a zip file containing php files.
#### Expected result
The file should upload.
#### Actual result
The file doesn't upload and not error message is shown. The file list just refreshes.
#### System information (as much as possible)
Tested on several 3.5.1 installations
#### Additional comments
I tried to upload the test zip files here but GitHub says "unfortunately, we don't support that file type", which is also pretty confusing as ZIP is listed as a supported file type, but it's way better than joomla's omission of an error whatsoever.
| non_priority | media manager not outputing any error when trying to upload a zip file that contains php files as far as i understand there is a new security feature that scans the content of a zip file uploaded through the media manager and forbids the upload if that zip file contains any php files in my case this is pretty stupid as i want people to upload zip files containing php code but anyway that s not the point of this issue here when uploading such an incorrect file i would expect to at least get an error message or at a warning instead the page just refreshes and no message is displayed steps to reproduce the issue go to media manager make sure that you have set the zip file extension in the allowed extension types try to upload a zip file containing php files expected result the file should upload actual result the file doesn t upload and not error message is shown the file list just refreshes system information as much as possible tested on several installations additional comments i tried to upload the test zip files here but github says unfortunately we don t support that file type which is also pretty confusing as zip is listed as a supported file type but it s way better than joomla s omission of an error whatsoever | 0 |
10,474 | 13,247,808,536 | IssuesEvent | 2020-08-19 17:52:54 | department-of-veterans-affairs/notification-api | https://api.github.com/repos/department-of-veterans-affairs/notification-api | opened | Approach with storing assets for VANotify | Process Task Release | Filip to add details on the options that were discussed.
*Notes:*
- The existing approach for storing VA assets needs to be resolved to work in Production
- Interim approach for storing assets is in card #145
- Sync with Filip on options that he has already discussed for the interim, and the limitations with the VA
- Document the decision on how/where to store assets for VANotify
| 1.0 | Approach with storing assets for VANotify - Filip to add details on the options that were discussed.
*Notes:*
- The existing approach for storing VA assets needs to be resolved to work in Production
- Interim approach for storing assets is in card #145
- Sync with Filip on options that he has already discussed for the interim, and the limitations with the VA
- Document the decision on how/where to store assets for VANotify
| non_priority | approach with storing assets for vanotify filip to add details on the options that were discussed notes the existing approach for storing va assets needs to be resolved to work in production interim approach for storing assets is in card sync with filip on options that he has already discussed for the interim and the limitations with the va document the decision on how where to store assets for vanotify | 0 |
267,497 | 28,509,065,179 | IssuesEvent | 2023-04-19 01:32:16 | dpteam/RK3188_TABLET | https://api.github.com/repos/dpteam/RK3188_TABLET | closed | CVE-2014-9322 (High) detected in linuxv3.0, linuxv3.5 - autoclosed | Mend: dependency security vulnerability | ## CVE-2014-9322 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.0</b>, <b>linuxv3.5</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
arch/x86/kernel/entry_64.S in the Linux kernel before 3.17.5 does not properly handle faults associated with the Stack Segment (SS) segment register, which allows local users to gain privileges by triggering an IRET instruction that leads to access to a GS Base address from the wrong space.
<p>Publish Date: 2014-12-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-9322>CVE-2014-9322</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-9322">https://nvd.nist.gov/vuln/detail/CVE-2014-9322</a></p>
<p>Release Date: 2014-12-17</p>
<p>Fix Resolution: 3.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2014-9322 (High) detected in linuxv3.0, linuxv3.5 - autoclosed - ## CVE-2014-9322 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.0</b>, <b>linuxv3.5</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
arch/x86/kernel/entry_64.S in the Linux kernel before 3.17.5 does not properly handle faults associated with the Stack Segment (SS) segment register, which allows local users to gain privileges by triggering an IRET instruction that leads to access to a GS Base address from the wrong space.
<p>Publish Date: 2014-12-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-9322>CVE-2014-9322</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-9322">https://nvd.nist.gov/vuln/detail/CVE-2014-9322</a></p>
<p>Release Date: 2014-12-17</p>
<p>Fix Resolution: 3.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in autoclosed cve high severity vulnerability vulnerable libraries vulnerability details arch kernel entry s in the linux kernel before does not properly handle faults associated with the stack segment ss segment register which allows local users to gain privileges by triggering an iret instruction that leads to access to a gs base address from the wrong space publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
216,294 | 16,655,474,864 | IssuesEvent | 2021-06-05 12:50:39 | mdl29/scratchy | https://api.github.com/repos/mdl29/scratchy | closed | [doc] better contributing.md | documentation | - [x] some link doesn't work or are in double
- [x] make pr look like issue section
- [ ] add legend on png
- [x] fix typo
- [x] some other thing | 1.0 | [doc] better contributing.md - - [x] some link doesn't work or are in double
- [x] make pr look like issue section
- [ ] add legend on png
- [x] fix typo
- [x] some other thing | non_priority | better contributing md some link doesn t work or are in double make pr look like issue section add legend on png fix typo some other thing | 0 |
31,976 | 7,471,233,414 | IssuesEvent | 2018-04-03 08:36:48 | Microsoft/vscode-arduino | https://api.github.com/repos/Microsoft/vscode-arduino | closed | Cannot set output folder to absolute path | P2 bug code ready | So to me it seems sensible to but the build output folder into the %temp% directory (usually `C:\Users\<user>\AppData\Local\Temp`). This is how the Arduino IDE does it as well.
However, setting the output variable in `arduino.json` will lead to an error message when starting the build and upload:
### arduino.json
```
{
"sketch": "bcm_data_direction.ino",
"board": "arduino:avr:micro",
"port": "COM4",
"output": "C:\\Users\\<user>\\AppData\\Local\\Temp\\ArduinoBuild"
}
```
### Output
```
[Starting] Upload sketch - bcm_data_direction.ino
Looking for library 'user32'
Adding paths from jna.library.path: null
Trying user32.dll
Found library 'user32' at user32.dll
Looking for library 'shell32'
Adding paths from jna.library.path: null
Trying shell32.dll
Found library 'shell32' at shell32.dll
Looking for library 'Ole32'
Adding paths from jna.library.path: null
Trying Ole32.dll
Found library 'Ole32' at Ole32.dll
Loading configuration...
Initialising packages...
Preparing boards...
Verifying...
[Error] Exit with code=1
```
I also tried this with other absolute paths, and even on different drives. I get this very same error message every single time.
It does work perfectly when setting a relative path. | 1.0 | Cannot set output folder to absolute path - So to me it seems sensible to but the build output folder into the %temp% directory (usually `C:\Users\<user>\AppData\Local\Temp`). This is how the Arduino IDE does it as well.
However, setting the output variable in `arduino.json` will lead to an error message when starting the build and upload:
### arduino.json
```
{
"sketch": "bcm_data_direction.ino",
"board": "arduino:avr:micro",
"port": "COM4",
"output": "C:\\Users\\<user>\\AppData\\Local\\Temp\\ArduinoBuild"
}
```
### Output
```
[Starting] Upload sketch - bcm_data_direction.ino
Looking for library 'user32'
Adding paths from jna.library.path: null
Trying user32.dll
Found library 'user32' at user32.dll
Looking for library 'shell32'
Adding paths from jna.library.path: null
Trying shell32.dll
Found library 'shell32' at shell32.dll
Looking for library 'Ole32'
Adding paths from jna.library.path: null
Trying Ole32.dll
Found library 'Ole32' at Ole32.dll
Loading configuration...
Initialising packages...
Preparing boards...
Verifying...
[Error] Exit with code=1
```
I also tried this with other absolute paths, and even on different drives. I get this very same error message every single time.
It does work perfectly when setting a relative path. | non_priority | cannot set output folder to absolute path so to me it seems sensible to but the build output folder into the temp directory usually c users appdata local temp this is how the arduino ide does it as well however setting the output variable in arduino json will lead to an error message when starting the build and upload arduino json sketch bcm data direction ino board arduino avr micro port output c users appdata local temp arduinobuild output upload sketch bcm data direction ino looking for library adding paths from jna library path null trying dll found library at dll looking for library adding paths from jna library path null trying dll found library at dll looking for library adding paths from jna library path null trying dll found library at dll loading configuration initialising packages preparing boards verifying exit with code i also tried this with other absolute paths and even on different drives i get this very same error message every single time it does work perfectly when setting a relative path | 0 |
16,445 | 4,053,937,855 | IssuesEvent | 2016-05-24 10:23:38 | ES-DOC/esdoc-docs | https://api.github.com/repos/ES-DOC/esdoc-docs | opened | External review for ScenarioMIP experiment documentation | CMIP6 Documentation | Review the ScenarioMIP experiments with the PIs. | 1.0 | External review for ScenarioMIP experiment documentation - Review the ScenarioMIP experiments with the PIs. | non_priority | external review for scenariomip experiment documentation review the scenariomip experiments with the pis | 0 |
12,520 | 7,895,867,324 | IssuesEvent | 2018-06-29 06:12:40 | Azure/azure-functions-durable-extension | https://api.github.com/repos/Azure/azure-functions-durable-extension | opened | Concurrency throttle implementation slows down processing unnecessarily | dtfx performance | Related to https://github.com/Azure/azure-functions-durable-extension/issues/368.
The Durable Task Framework dispatcher throttling logic currently *sleeps* the dispatcher for 5 seconds every time it reaches the configured number of max concurrent work items. This 5 second backoff can have a significant impact on throughput because it effectively halts all queue processing for those 5 seconds.
The impact is expected to be even greater for orchestrator message processing because of how rapidly new control queue work items can be scheduled, thus causing the throttle limits to be hit very easily, especially after all activity functions have finished executing. The theory is that this results in very inefficient use of CPU cycles on the VM because a worker VM, for example, may only process one batch of messages every 5 seconds even if it could theoretically process them much faster.
The workaround is to create very high concurrency limits so that the throttling logic will not kick in. | True | Concurrency throttle implementation slows down processing unnecessarily - Related to https://github.com/Azure/azure-functions-durable-extension/issues/368.
The Durable Task Framework dispatcher throttling logic currently *sleeps* the dispatcher for 5 seconds every time it reaches the configured number of max concurrent work items. This 5 second backoff can have a significant impact on throughput because it effectively halts all queue processing for those 5 seconds.
The impact is expected to be even greater for orchestrator message processing because of how rapidly new control queue work items can be scheduled, thus causing the throttle limits to be hit very easily, especially after all activity functions have finished executing. The theory is that this results in very inefficient use of CPU cycles on the VM because a worker VM, for example, may only process one batch of messages every 5 seconds even if it could theoretically process them much faster.
The workaround is to create very high concurrency limits so that the throttling logic will not kick in. | non_priority | concurrency throttle implementation slows down processing unnecessarily related to the durable task framework dispatcher throttling logic currently sleeps the dispatcher for seconds every time it reaches the configured number of max concurrent work items this second backoff can have a significant impact on throughput because it effectively halts all queue processing for those seconds the impact is expected to be even greater for orchestrator message processing because of how rapidly new control queue work items can be scheduled thus causing the throttle limits to be hit very easily especially after all activity functions have finished executing the theory is that this results in very inefficient use of cpu cycles on the vm because a worker vm for example may only process one batch of messages every seconds even if it could theoretically process them much faster the workaround is to create very high concurrency limits so that the throttling logic will not kick in | 0 |
48,847 | 25,860,024,313 | IssuesEvent | 2022-12-13 16:17:55 | mgerhardy/vengi | https://api.github.com/repos/mgerhardy/vengi | closed | VOXEDIT: when rendering all viewports there is a massive slowdown | bug voxedit animation performance | re-arranging the viewports slows down voxedit.

| True | VOXEDIT: when rendering all viewports there is a massive slowdown - re-arranging the viewports slows down voxedit.

| non_priority | voxedit when rendering all viewports there is a massive slowdown re arranging the viewports slows down voxedit | 0 |
126,026 | 26,769,756,135 | IssuesEvent | 2023-01-31 13:19:15 | Epicalable/JARVIS | https://api.github.com/repos/Epicalable/JARVIS | closed | Flight Tracker Issue Tracker | Code Bug Code Enhancement | Currently this function is in beta and is currently under testing.
We will also be trying to create a GUI to show the data rather than on the mainframe terminal.
R&D and Coding Department will help to oversee this function's issues and will close this issue once all problems are rectified and solved. | 2.0 | Flight Tracker Issue Tracker - Currently this function is in beta and is currently under testing.
We will also be trying to create a GUI to show the data rather than on the mainframe terminal.
R&D and Coding Department will help to oversee this function's issues and will close this issue once all problems are rectified and solved. | non_priority | flight tracker issue tracker currently this function is in beta and is currently under testing we will also be trying to create a gui to show the data rather than on the mainframe terminal r d and coding department will help to oversee this function s issues and will close this issue once all problems are rectified and solved | 0 |
405,248 | 27,509,201,355 | IssuesEvent | 2023-03-06 07:21:14 | llvm/llvm-project | https://api.github.com/repos/llvm/llvm-project | closed | [libc++] Mark `split_view` as finished | documentation libc++ release:backport | This is to ensure the status notes in release 16 are up-to-date. | 1.0 | [libc++] Mark `split_view` as finished - This is to ensure the status notes in release 16 are up-to-date. | non_priority | mark split view as finished this is to ensure the status notes in release are up to date | 0 |
4,750 | 17,376,620,575 | IssuesEvent | 2021-07-30 22:49:16 | CDCgov/prime-field-teams | https://api.github.com/repos/CDCgov/prime-field-teams | closed | Precision Diagnostics Onboarding | Epic Psyche State-Louisiana sender-automation | **_Standard Tasks / Onboarding Steps:_**
**Kickoff Meetings/Follow Up:**
**Solution Planning:**
- Mapping
- Brainstorming/Grooming
- Complete RS Solution Feature Request (if needed)
**Solution Development:**
- RS Schema
- Other Dev (VBScript, etc)
**Solution Testing:**
- Staging
- Prod/Test
- RS Feature Request (if needed)
- Sender Acceptance w/ Client Signoff
**Solution Implementation:**
- Plan & Schedule Go Live
- GO LIVE and Post LIVE Review | 1.0 | Precision Diagnostics Onboarding - **_Standard Tasks / Onboarding Steps:_**
**Kickoff Meetings/Follow Up:**
**Solution Planning:**
- Mapping
- Brainstorming/Grooming
- Complete RS Solution Feature Request (if needed)
**Solution Development:**
- RS Schema
- Other Dev (VBScript, etc)
**Solution Testing:**
- Staging
- Prod/Test
- RS Feature Request (if needed)
- Sender Acceptance w/ Client Signoff
**Solution Implementation:**
- Plan & Schedule Go Live
- GO LIVE and Post LIVE Review | non_priority | precision diagnostics onboarding standard tasks onboarding steps kickoff meetings follow up solution planning mapping brainstorming grooming complete rs solution feature request if needed solution development rs schema other dev vbscript etc solution testing staging prod test rs feature request if needed sender acceptance w client signoff solution implementation plan schedule go live go live and post live review | 0 |
32,882 | 7,612,629,223 | IssuesEvent | 2018-05-01 18:14:57 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Hard coupling between com_users and plg_user_joomla | No Code Attached Yet | Within [`UsersModelUser::getForm()`](https://github.com/joomla/joomla-cms/blob/49e681fc66ecdd76388fa9e9cb3a92986c075e69/administrator/components/com_users/models/user.php#L131-L138) there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin. Either this capability should be fully moved to the plugin (meaning the form alterations should happen via plugin event) or be fully moved to the component (meaning the mail sending code exists in the component). | 1.0 | Hard coupling between com_users and plg_user_joomla - Within [`UsersModelUser::getForm()`](https://github.com/joomla/joomla-cms/blob/49e681fc66ecdd76388fa9e9cb3a92986c075e69/administrator/components/com_users/models/user.php#L131-L138) there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin. Either this capability should be fully moved to the plugin (meaning the form alterations should happen via plugin event) or be fully moved to the component (meaning the mail sending code exists in the component). | non_priority | hard coupling between com users and plg user joomla within there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin either this capability should be fully moved to the plugin meaning the form alterations should happen via plugin event or be fully moved to the component meaning the mail sending code exists in the component | 0 |
104,666 | 11,418,274,188 | IssuesEvent | 2020-02-03 03:51:20 | consento-org/consento-website | https://api.github.com/repos/consento-org/consento-website | closed | Add H2020 funding note | documentation | Add the following sentence to the project website:
> This project has received funding from the European Union’s Horizon 2020
> research and innovation programme within the framework of the LEDGER
> Project funded under grant agreement No825268.
Together with the EU emblem and LEDGER logo:

(this is just a screenshot, use original logos with appropriate quality) | 1.0 | Add H2020 funding note - Add the following sentence to the project website:
> This project has received funding from the European Union’s Horizon 2020
> research and innovation programme within the framework of the LEDGER
> Project funded under grant agreement No825268.
Together with the EU emblem and LEDGER logo:

(this is just a screenshot, use original logos with appropriate quality) | non_priority | add funding note add the following sentence to the project website this project has received funding from the european union’s horizon research and innovation programme within the framework of the ledger project funded under grant agreement together with the eu emblem and ledger logo this is just a screenshot use original logos with appropriate quality | 0 |
277,032 | 30,594,516,355 | IssuesEvent | 2023-07-21 20:23:06 | justunsix/automatetheboringstuff-py-tests | https://api.github.com/repos/justunsix/automatetheboringstuff-py-tests | opened | CVE-2023-25663 (High) detected in tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl | Mend: dependency security vulnerability | ## CVE-2023-25663 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/42/24/830571895f0927fe205a23309b136520c7914921420bd1e81aff1da47bb1/tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/42/24/830571895f0927fe205a23309b136520c7914921420bd1e81aff1da47bb1/tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /src/project/data-science/requirements.txt</p>
<p>Path to vulnerable library: /src/project/data-science/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/justunsix/automatetheboringstuff-py-tests/commit/92e57f9e81da15812523bf929f8ad33bdae5e967">92e57f9e81da15812523bf929f8ad33bdae5e967</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, when `ctx->step_containter()` is a null ptr, the Lookup function will be executed with a null pointer. A fix is included in TensorFlow 2.12.0 and 2.11.1.
<p>Publish Date: 2023-03-25
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25663>CVE-2023-25663</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-64jg-wjww-7c5w">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-64jg-wjww-7c5w</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-25663 (High) detected in tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl - ## CVE-2023-25663 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/42/24/830571895f0927fe205a23309b136520c7914921420bd1e81aff1da47bb1/tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/42/24/830571895f0927fe205a23309b136520c7914921420bd1e81aff1da47bb1/tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /src/project/data-science/requirements.txt</p>
<p>Path to vulnerable library: /src/project/data-science/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.11.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/justunsix/automatetheboringstuff-py-tests/commit/92e57f9e81da15812523bf929f8ad33bdae5e967">92e57f9e81da15812523bf929f8ad33bdae5e967</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, when `ctx->step_containter()` is a null ptr, the Lookup function will be executed with a null pointer. A fix is included in TensorFlow 2.12.0 and 2.11.1.
<p>Publish Date: 2023-03-25
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25663>CVE-2023-25663</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-64jg-wjww-7c5w">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-64jg-wjww-7c5w</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tensorflow manylinux whl cve high severity vulnerability vulnerable library tensorflow manylinux whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file src project data science requirements txt path to vulnerable library src project data science requirements txt dependency hierarchy x tensorflow manylinux whl vulnerable library found in head commit a href found in base branch main vulnerability details tensorflow is an open source platform for machine learning prior to versions and when ctx step containter is a null ptr the lookup function will be executed with a null pointer a fix is included in tensorflow and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend | 0 |
64,951 | 12,507,457,527 | IssuesEvent | 2020-06-02 14:09:25 | Regalis11/Barotrauma | https://api.github.com/repos/Regalis11/Barotrauma | closed | [v0.1000.0.0 Unstable] Error while trying to load a save | Bug Code | - [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
A previously saved game won't load. Error message starts with:
```
Loading save <path> failed (Object reference not set to an instance of an object.)
at Barotrauma.Job..ctor(Xelement element) in
C:\Users\Joonas\Documents\Barotrauma-development\Barotrauma\BarotraumaShared\SharedSource\Characters\Jobs\Job.cs:line 74
```
**Steps To Reproduce**
1. Start a campaign.
2. Hire or spawn a character with the job Assistant or any other (custom) job with no skill defined as primary in the xml.
3. Save the game.
4. Try to load it again.
**Version**
0.1000.0.0 Unstable
| 1.0 | [v0.1000.0.0 Unstable] Error while trying to load a save - - [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
A previously saved game won't load. Error message starts with:
```
Loading save <path> failed (Object reference not set to an instance of an object.)
at Barotrauma.Job..ctor(Xelement element) in
C:\Users\Joonas\Documents\Barotrauma-development\Barotrauma\BarotraumaShared\SharedSource\Characters\Jobs\Job.cs:line 74
```
**Steps To Reproduce**
1. Start a campaign.
2. Hire or spawn a character with the job Assistant or any other (custom) job with no skill defined as primary in the xml.
3. Save the game.
4. Try to load it again.
**Version**
0.1000.0.0 Unstable
| non_priority | error while trying to load a save i have searched the issue tracker to check if the issue has already been reported description a previously saved game won t load error message starts with loading save failed object reference not set to an instance of an object at barotrauma job ctor xelement element in c users joonas documents barotrauma development barotrauma barotraumashared sharedsource characters jobs job cs line steps to reproduce start a campaign hire or spawn a character with the job assistant or any other custom job with no skill defined as primary in the xml save the game try to load it again version unstable | 0 |
32,387 | 15,364,259,636 | IssuesEvent | 2021-03-01 21:45:00 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | AutoPublishAlias makes sam deployments slow | area/performance | We have added AutoPublishAlias: live to help preprovisioning our specific functions for warm up.
```
AutoPublishAlias: live
ProvisionedConcurrencyConfig:
ProvisionedConcurrentExecutions: 15
```
**Description:**
When we use sam deploy we are finding the above section consumes lot of time to finish deployment compared to without the above section. We understand that this is due pre warming of the lambdas for preprovisioning size requested (in this case 15). But, is there a way to avoid blocking the sam deploy from waiting for the lambdas to come up. This will drastically improve our deployment and development velocity.
**Steps to reproduce the issue:**
1. Add the section above
2. sam deploy takes in the order of 15 mins or more for deployment.
**Observed result:**
**Expected result:**
Avoid paying deployment time penalty to add the ProvisionedConcurrencyConfig section to the function definition.
| True | AutoPublishAlias makes sam deployments slow - We have added AutoPublishAlias: live to help preprovisioning our specific functions for warm up.
```
AutoPublishAlias: live
ProvisionedConcurrencyConfig:
ProvisionedConcurrentExecutions: 15
```
**Description:**
When we use sam deploy we are finding the above section consumes lot of time to finish deployment compared to without the above section. We understand that this is due pre warming of the lambdas for preprovisioning size requested (in this case 15). But, is there a way to avoid blocking the sam deploy from waiting for the lambdas to come up. This will drastically improve our deployment and development velocity.
**Steps to reproduce the issue:**
1. Add the section above
2. sam deploy takes in the order of 15 mins or more for deployment.
**Observed result:**
**Expected result:**
Avoid paying deployment time penalty to add the ProvisionedConcurrencyConfig section to the function definition.
| non_priority | autopublishalias makes sam deployments slow we have added autopublishalias live to help preprovisioning our specific functions for warm up autopublishalias live provisionedconcurrencyconfig provisionedconcurrentexecutions description when we use sam deploy we are finding the above section consumes lot of time to finish deployment compared to without the above section we understand that this is due pre warming of the lambdas for preprovisioning size requested in this case but is there a way to avoid blocking the sam deploy from waiting for the lambdas to come up this will drastically improve our deployment and development velocity steps to reproduce the issue add the section above sam deploy takes in the order of mins or more for deployment observed result expected result avoid paying deployment time penalty to add the provisionedconcurrencyconfig section to the function definition | 0 |
24,548 | 11,046,988,902 | IssuesEvent | 2019-12-09 17:59:24 | LevyForchh/webdataconnector | https://api.github.com/repos/LevyForchh/webdataconnector | opened | CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js | security vulnerability | ## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/webdataconnector/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: /webdataconnector/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/webdataconnector/commit/5e5d2c912c2a10aa798a92712142633ee3a0580f">5e5d2c912c2a10aa798a92712142633ee3a0580f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> --> | True | CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - ## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/webdataconnector/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: /webdataconnector/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/webdataconnector/commit/5e5d2c912c2a10aa798a92712142633ee3a0580f">5e5d2c912c2a10aa798a92712142633ee3a0580f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm webdataconnector node modules selenium webdriver lib test data droppableitems html path to vulnerable library webdataconnector node modules selenium webdriver lib test data js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag vulnerabilityurl | 0 |
44,320 | 11,426,837,309 | IssuesEvent | 2020-02-03 22:55:15 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | opened | Support gradle run command for multi-node cluster | :Core/Infra/Build >enhancement v8.0.0 | The command `./gradlew run` works great for single-node cluster. But there is no pre-configured way to launch a multi-node cluster. Following suggestion by @mark-vieira, I added `numberOfNodes = 3` after [this line](https://github.com/elastic/elasticsearch/blob/master/distribution/build.gradle#L421). However the cluster still fails to form. The error message shows the nodes could **not** find the expected `unicast_hosts.txt` file and hence could not elect master node.
When manually add the above missing file for one of the nodes, the cluster formed immediately afterwards. It would be great to have these manual steps supported out of the box for the run task. Thanks. | 1.0 | Support gradle run command for multi-node cluster - The command `./gradlew run` works great for single-node cluster. But there is no pre-configured way to launch a multi-node cluster. Following suggestion by @mark-vieira, I added `numberOfNodes = 3` after [this line](https://github.com/elastic/elasticsearch/blob/master/distribution/build.gradle#L421). However the cluster still fails to form. The error message shows the nodes could **not** find the expected `unicast_hosts.txt` file and hence could not elect master node.
When manually add the above missing file for one of the nodes, the cluster formed immediately afterwards. It would be great to have these manual steps supported out of the box for the run task. Thanks. | non_priority | support gradle run command for multi node cluster the command gradlew run works great for single node cluster but there is no pre configured way to launch a multi node cluster following suggestion by mark vieira i added numberofnodes after however the cluster still fails to form the error message shows the nodes could not find the expected unicast hosts txt file and hence could not elect master node when manually add the above missing file for one of the nodes the cluster formed immediately afterwards it would be great to have these manual steps supported out of the box for the run task thanks | 0 |
109,852 | 11,660,012,471 | IssuesEvent | 2020-03-03 01:54:10 | personium/new-docs | https://api.github.com/repos/personium/new-docs | closed | latest personium.io is not reflected | documentation | When should personium.io be deprecated?
[latest update](https://github.com/personium/personium.io/commit/7f78864e26325d50a5cee901bec825b62f894ea2)
## Old slack invitation
- [Japanese](https://github.com/personium/new-docs/blob/master/website/static/ja/index.html#L464)
- [English](https://github.com/personium/new-docs/blob/master/website/static/en/index.html#L455) | 1.0 | latest personium.io is not reflected - When should personium.io be deprecated?
[latest update](https://github.com/personium/personium.io/commit/7f78864e26325d50a5cee901bec825b62f894ea2)
## Old slack invitation
- [Japanese](https://github.com/personium/new-docs/blob/master/website/static/ja/index.html#L464)
- [English](https://github.com/personium/new-docs/blob/master/website/static/en/index.html#L455) | non_priority | latest personium io is not reflected when should personium io be deprecated old slack invitation | 0 |
34,490 | 16,583,087,207 | IssuesEvent | 2021-05-31 14:27:11 | microsoft/reverse-proxy | https://api.github.com/repos/microsoft/reverse-proxy | closed | Measure performance impact of passive health checks | Type: Idea tenet-performance | We need to understand what is the overhead of having passive health checks enabled. One of suspicious places to take closer look is in [TransportFailureRateHealthPolicy.EvaluateProxiedRequest](https://github.com/microsoft/reverse-proxy/blob/51b3a7df9f54f848d6ee9028dc5a7cfb9ee48d02/src/ReverseProxy/Service/HealthChecks/TransportFailureRateHealthPolicy.cs#L60) which takes destination-scoped lock on each request. | True | Measure performance impact of passive health checks - We need to understand what is the overhead of having passive health checks enabled. One of suspicious places to take closer look is in [TransportFailureRateHealthPolicy.EvaluateProxiedRequest](https://github.com/microsoft/reverse-proxy/blob/51b3a7df9f54f848d6ee9028dc5a7cfb9ee48d02/src/ReverseProxy/Service/HealthChecks/TransportFailureRateHealthPolicy.cs#L60) which takes destination-scoped lock on each request. | non_priority | measure performance impact of passive health checks we need to understand what is the overhead of having passive health checks enabled one of suspicious places to take closer look is in which takes destination scoped lock on each request | 0 |
75,887 | 14,532,689,666 | IssuesEvent | 2020-12-14 22:53:39 | Everything-Explained/web-client | https://api.github.com/repos/Everything-Explained/web-client | closed | Use Vue 3 composition API everywhere | chore:clean-code chore:refactor scope:engineering | I've had this love-hate relationships with functional programming, but it's definitely a much more organized way to program. Having your code laid out in the order that it executes, especially with a larger component, is a god-send.
Each Class that I move over into the new project, should be converted to this composition style where every operation is in a logical order. | 1.0 | Use Vue 3 composition API everywhere - I've had this love-hate relationships with functional programming, but it's definitely a much more organized way to program. Having your code laid out in the order that it executes, especially with a larger component, is a god-send.
Each Class that I move over into the new project, should be converted to this composition style where every operation is in a logical order. | non_priority | use vue composition api everywhere i ve had this love hate relationships with functional programming but it s definitely a much more organized way to program having your code laid out in the order that it executes especially with a larger component is a god send each class that i move over into the new project should be converted to this composition style where every operation is in a logical order | 0 |
3,528 | 2,770,963,874 | IssuesEvent | 2015-05-01 18:15:01 | jayspur11/NatCompHW3 | https://api.github.com/repos/jayspur11/NatCompHW3 | closed | Slides, question 2 | Documentation Finished Problem Finished | Reproduce patterns theta, lambda, mu, and alpha in the Gray-Scott Model (CA). You don't need to follow their color scheme. | 1.0 | Slides, question 2 - Reproduce patterns theta, lambda, mu, and alpha in the Gray-Scott Model (CA). You don't need to follow their color scheme. | non_priority | slides question reproduce patterns theta lambda mu and alpha in the gray scott model ca you don t need to follow their color scheme | 0 |
21,814 | 17,793,741,635 | IssuesEvent | 2021-08-31 19:23:33 | tailscale/tailscale | https://api.github.com/repos/tailscale/tailscale | closed | synology: upgrading the tailscale package from DSM6 to DSM7 doesn't copy over state | L3 Some users P3 Can't get started T6 Major usability OS-synology | When I upgraded my DSM6 synology to DSM7, Tailscale wouldn't start and had to be upgraded to the DSM7 build.
However, the state wasn't copied over resulting in an unregistered node.
I was able to manually copy over state by doing the following:
1. Stop tailscale via the UI
2. SSH into the machine
3. Run
```
sudo cp /var/packages/Tailscale/etc/* /var/packages/Tailscale/var/
sudo chown -R tailscale:tailscale /var/packages/Tailscale/var
```
4. Start tailscale from the UI | True | synology: upgrading the tailscale package from DSM6 to DSM7 doesn't copy over state - When I upgraded my DSM6 synology to DSM7, Tailscale wouldn't start and had to be upgraded to the DSM7 build.
However, the state wasn't copied over resulting in an unregistered node.
I was able to manually copy over state by doing the following:
1. Stop tailscale via the UI
2. SSH into the machine
3. Run
```
sudo cp /var/packages/Tailscale/etc/* /var/packages/Tailscale/var/
sudo chown -R tailscale:tailscale /var/packages/Tailscale/var
```
4. Start tailscale from the UI | non_priority | synology upgrading the tailscale package from to doesn t copy over state when i upgraded my synology to tailscale wouldn t start and had to be upgraded to the build however the state wasn t copied over resulting in an unregistered node i was able to manually copy over state by doing the following stop tailscale via the ui ssh into the machine run sudo cp var packages tailscale etc var packages tailscale var sudo chown r tailscale tailscale var packages tailscale var start tailscale from the ui | 0 |
226,020 | 24,931,678,210 | IssuesEvent | 2022-10-31 12:10:46 | lukebrogan-mend/railsgoat | https://api.github.com/repos/lukebrogan-mend/railsgoat | reopened | jquery.validate-1.17.0.min.js: 1 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.validate-1.17.0.min.js</b></p></summary>
<p>Client-side form validation made easy</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js</a></p>
<p>Path to vulnerable library: /app/assets/javascripts/jquery.validate.min.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/railsgoat/commit/10b00bba83d518df58b71c164d0be7c229d4b799">10b00bba83d518df58b71c164d0be7c229d4b799</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.validate version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-21252](https://www.mend.io/vulnerability-database/CVE-2021-21252) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jquery.validate-1.17.0.min.js | Direct | jquery-validation - 1.19.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-21252</summary>
### Vulnerable Library - <b>jquery.validate-1.17.0.min.js</b></p>
<p>Client-side form validation made easy</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js</a></p>
<p>Path to vulnerable library: /app/assets/javascripts/jquery.validate.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.validate-1.17.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/railsgoat/commit/10b00bba83d518df58b71c164d0be7c229d4b799">10b00bba83d518df58b71c164d0be7c229d4b799</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The jQuery Validation Plugin provides drop-in validation for your existing forms. It is published as an npm package "jquery-validation". jquery-validation before version 1.19.3 contains one or more regular expressions that are vulnerable to ReDoS (Regular Expression Denial of Service). This is fixed in 1.19.3.
<p>Publish Date: 2021-01-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21252>CVE-2021-21252</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jxwx-85vp-gvwm">https://github.com/advisories/GHSA-jxwx-85vp-gvwm</a></p>
<p>Release Date: 2021-01-13</p>
<p>Fix Resolution: jquery-validation - 1.19.3</p>
</p>
<p></p>
</details> | True | jquery.validate-1.17.0.min.js: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.validate-1.17.0.min.js</b></p></summary>
<p>Client-side form validation made easy</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js</a></p>
<p>Path to vulnerable library: /app/assets/javascripts/jquery.validate.min.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/railsgoat/commit/10b00bba83d518df58b71c164d0be7c229d4b799">10b00bba83d518df58b71c164d0be7c229d4b799</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.validate version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-21252](https://www.mend.io/vulnerability-database/CVE-2021-21252) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jquery.validate-1.17.0.min.js | Direct | jquery-validation - 1.19.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-21252</summary>
### Vulnerable Library - <b>jquery.validate-1.17.0.min.js</b></p>
<p>Client-side form validation made easy</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-validate/1.17.0/jquery.validate.min.js</a></p>
<p>Path to vulnerable library: /app/assets/javascripts/jquery.validate.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.validate-1.17.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/railsgoat/commit/10b00bba83d518df58b71c164d0be7c229d4b799">10b00bba83d518df58b71c164d0be7c229d4b799</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The jQuery Validation Plugin provides drop-in validation for your existing forms. It is published as an npm package "jquery-validation". jquery-validation before version 1.19.3 contains one or more regular expressions that are vulnerable to ReDoS (Regular Expression Denial of Service). This is fixed in 1.19.3.
<p>Publish Date: 2021-01-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21252>CVE-2021-21252</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jxwx-85vp-gvwm">https://github.com/advisories/GHSA-jxwx-85vp-gvwm</a></p>
<p>Release Date: 2021-01-13</p>
<p>Fix Resolution: jquery-validation - 1.19.3</p>
</p>
<p></p>
</details> | non_priority | jquery validate min js vulnerabilities highest severity is vulnerable library jquery validate min js client side form validation made easy library home page a href path to vulnerable library app assets javascripts jquery validate min js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery validate version remediation available high jquery validate min js direct jquery validation details cve vulnerable library jquery validate min js client side form validation made easy library home page a href path to vulnerable library app assets javascripts jquery validate min js dependency hierarchy x jquery validate min js vulnerable library found in head commit a href found in base branch master vulnerability details the jquery validation plugin provides drop in validation for your existing forms it is published as an npm package jquery validation jquery validation before version contains one or more regular expressions that are vulnerable to redos regular expression denial of service this is fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery validation | 0 |
80,135 | 23,124,531,245 | IssuesEvent | 2022-07-28 03:16:54 | jjik-muk/sikdorak | https://api.github.com/repos/jjik-muk/sikdorak | closed | [BE/build] API 문서화 설정 구성 | 🤖 BE 🏗️ build 📝 docs | ## 📝 구현 목록
- [x] Spring Restdocs 설정 추가
- [x] Swagger UI 설정 추가
- [x] Restdocs + Swagger UI 연동 설정 추가
<br>
## 💡 참고자료
- [Generate Swagger UI from Spring REST Docs](https://blog.jdriven.com/2021/10/generate-swagger-ui-from-spring-rest-docs/)
| 1.0 | [BE/build] API 문서화 설정 구성 - ## 📝 구현 목록
- [x] Spring Restdocs 설정 추가
- [x] Swagger UI 설정 추가
- [x] Restdocs + Swagger UI 연동 설정 추가
<br>
## 💡 참고자료
- [Generate Swagger UI from Spring REST Docs](https://blog.jdriven.com/2021/10/generate-swagger-ui-from-spring-rest-docs/)
| non_priority | api 문서화 설정 구성 📝 구현 목록 spring restdocs 설정 추가 swagger ui 설정 추가 restdocs swagger ui 연동 설정 추가 💡 참고자료 | 0 |
101,615 | 11,253,468,151 | IssuesEvent | 2020-01-11 16:29:20 | DCAFEngineering/dcaf_case_management | https://api.github.com/repos/DCAFEngineering/dcaf_case_management | closed | Add note to data dictionary about why archived patient data is null before 2019-10-21 | beginner friendly documentation | This is a good ticket if you are learning Github's pull request flow!
## Context
https://codefordc.slack.com/archives/C0E6APB36/p1571703212151600
We had a bug in the archive & export code leading to all patients being marked as having special circumstances (hint: 1/0 are not truthy/falsey, they're both truthy :sob:).
Because of this we bug we had every archived patient flagged as having special circumstances. We decided to mark all those patients as not having, instead, since we have no way of discovering this after archiving.
## Work to do
Add this note to the [Data Dictionary](https://github.com/DCAFEngineering/dcaf_case_management/blob/master/docs/DATA_DICTIONARY.md):
> Archived Patients prior to 2019-10-21 will always have `had_special_circumstances` set to null due to a bug in the archiving processing not preserving the value. Archived patients after that date correctly reflect whether or not they had special circumstances. See [issue #1836](https://github.com/DCAFEngineering/dcaf_case_management/issues/1836) for further context. | 1.0 | Add note to data dictionary about why archived patient data is null before 2019-10-21 - This is a good ticket if you are learning Github's pull request flow!
## Context
https://codefordc.slack.com/archives/C0E6APB36/p1571703212151600
We had a bug in the archive & export code leading to all patients being marked as having special circumstances (hint: 1/0 are not truthy/falsey, they're both truthy :sob:).
Because of this we bug we had every archived patient flagged as having special circumstances. We decided to mark all those patients as not having, instead, since we have no way of discovering this after archiving.
## Work to do
Add this note to the [Data Dictionary](https://github.com/DCAFEngineering/dcaf_case_management/blob/master/docs/DATA_DICTIONARY.md):
> Archived Patients prior to 2019-10-21 will always have `had_special_circumstances` set to null due to a bug in the archiving processing not preserving the value. Archived patients after that date correctly reflect whether or not they had special circumstances. See [issue #1836](https://github.com/DCAFEngineering/dcaf_case_management/issues/1836) for further context. | non_priority | add note to data dictionary about why archived patient data is null before this is a good ticket if you are learning github s pull request flow context we had a bug in the archive export code leading to all patients being marked as having special circumstances hint are not truthy falsey they re both truthy sob because of this we bug we had every archived patient flagged as having special circumstances we decided to mark all those patients as not having instead since we have no way of discovering this after archiving work to do add this note to the archived patients prior to will always have had special circumstances set to null due to a bug in the archiving processing not preserving the value archived patients after that date correctly reflect whether or not they had special circumstances see for further context | 0 |
94,941 | 8,527,040,855 | IssuesEvent | 2018-11-02 18:10:18 | ukwa/ukwa-ui | https://api.github.com/repos/ukwa/ukwa-ui | closed | Spelling mistake in a Collection Summary | Text testing: ready to test | "adminstered"

For some reason this appears to be hardcoded! #193 raised to clarify whether that is the case. It's possible it is just a data fix though, and the hardcoding is a red herring from an earlier, leftover data load. | 2.0 | Spelling mistake in a Collection Summary - "adminstered"

For some reason this appears to be hardcoded! #193 raised to clarify whether that is the case. It's possible it is just a data fix though, and the hardcoding is a red herring from an earlier, leftover data load. | non_priority | spelling mistake in a collection summary adminstered for some reason this appears to be hardcoded raised to clarify whether that is the case it s possible it is just a data fix though and the hardcoding is a red herring from an earlier leftover data load | 0 |
4,138 | 19,663,538,793 | IssuesEvent | 2022-01-10 19:39:56 | VA-Explorer/va_explorer | https://api.github.com/repos/VA-Explorer/va_explorer | opened | Consider ways to prevent unnecessary API calls via clientside callbacks | Type: Maintainance | **What is the expected state?**
Clientside callbacks are utilized in the context described in #173
**What is the actual state?**
No optimization
**Relevant context**
`va_explorer/va_analytics/dash_apps/va_dashboard.py`
| True | Consider ways to prevent unnecessary API calls via clientside callbacks - **What is the expected state?**
Clientside callbacks are utilized in the context described in #173
**What is the actual state?**
No optimization
**Relevant context**
`va_explorer/va_analytics/dash_apps/va_dashboard.py`
| non_priority | consider ways to prevent unnecessary api calls via clientside callbacks what is the expected state clientside callbacks are utilized in the context described in what is the actual state no optimization relevant context va explorer va analytics dash apps va dashboard py | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.