Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
299,570 | 9,205,621,911 | IssuesEvent | 2019-03-08 11:07:15 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | Grayscale Band Min Max Values Not Saved | Category: Rasters Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **David A- Riggs -** (David A- Riggs -)
Original Redmine Issue: 945, https://issues.qgis.org/issues/945
Original Assignee: ersts -
---
1. Add a new raster layer, grayscale .tiff (DEM from ftp://ftp.wvgis.wvu.edu/pub/Clearinghouse/DEM/3m/Tiff/ )
2. Edit Properties: Select Grayscale Band Scaling "Custom Min Max Value", set values for Min and Max. The color scaling is adjusted as expected.
3. Save as new project. Close QGIS. Open QGIS. Open project.
Result: the .tiff file as been set back to Grayscale Band Scaling "Std Deviation", Min and Max values have reset to actual values in file. In other words, all custom Grayscale Band Scaling settings have been lost.
Expected behavior: "Custom Min Max Value" setting persisted, Min and Max values persisted.
This was tested with both HEAD/trunk and with the 0.9.2rc1 Debian packages.
| 1.0 | Grayscale Band Min Max Values Not Saved - ---
Author Name: **David A- Riggs -** (David A- Riggs -)
Original Redmine Issue: 945, https://issues.qgis.org/issues/945
Original Assignee: ersts -
---
1. Add a new raster layer, grayscale .tiff (DEM from ftp://ftp.wvgis.wvu.edu/pub/Clearinghouse/DEM/3m/Tiff/ )
2. Edit Properties: Select Grayscale Band Scaling "Custom Min Max Value", set values for Min and Max. The color scaling is adjusted as expected.
3. Save as new project. Close QGIS. Open QGIS. Open project.
Result: the .tiff file as been set back to Grayscale Band Scaling "Std Deviation", Min and Max values have reset to actual values in file. In other words, all custom Grayscale Band Scaling settings have been lost.
Expected behavior: "Custom Min Max Value" setting persisted, Min and Max values persisted.
This was tested with both HEAD/trunk and with the 0.9.2rc1 Debian packages.
| priority | grayscale band min max values not saved author name david a riggs david a riggs original redmine issue original assignee ersts add a new raster layer grayscale tiff dem from ftp ftp wvgis wvu edu pub clearinghouse dem tiff edit properties select grayscale band scaling custom min max value set values for min and max the color scaling is adjusted as expected save as new project close qgis open qgis open project result the tiff file as been set back to grayscale band scaling std deviation min and max values have reset to actual values in file in other words all custom grayscale band scaling settings have been lost expected behavior custom min max value setting persisted min and max values persisted this was tested with both head trunk and with the debian packages | 1 |
281,985 | 30,889,065,524 | IssuesEvent | 2023-08-04 02:11:23 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2018-5873 (High) detected in linux-yocto-devv4.2.8 | Mend: dependency security vulnerability | ## CVE-2018-5873 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nsfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nsfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the __ns_get_path function in fs/nsfs.c in the Linux kernel before 4.11. Due to a race condition when accessing files, a Use After Free condition can occur. This also affects all Android releases from CAF using the Linux kernel (Android for MSM, Firefox OS for MSM, QRD Android) before security patch level 2018-07-05.
<p>Publish Date: 2018-07-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5873>CVE-2018-5873</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2018-5873">https://www.linuxkernelcves.com/cves/CVE-2018-5873</a></p>
<p>Release Date: 2018-07-06</p>
<p>Fix Resolution: v4.1.50,v4.4.116,v4.9.82,v4.11-rc8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-5873 (High) detected in linux-yocto-devv4.2.8 - ## CVE-2018-5873 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nsfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nsfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the __ns_get_path function in fs/nsfs.c in the Linux kernel before 4.11. Due to a race condition when accessing files, a Use After Free condition can occur. This also affects all Android releases from CAF using the Linux kernel (Android for MSM, Firefox OS for MSM, QRD Android) before security patch level 2018-07-05.
<p>Publish Date: 2018-07-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5873>CVE-2018-5873</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2018-5873">https://www.linuxkernelcves.com/cves/CVE-2018-5873</a></p>
<p>Release Date: 2018-07-06</p>
<p>Fix Resolution: v4.1.50,v4.4.116,v4.9.82,v4.11-rc8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linux yocto cve high severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files fs nsfs c fs nsfs c vulnerability details an issue was discovered in the ns get path function in fs nsfs c in the linux kernel before due to a race condition when accessing files a use after free condition can occur this also affects all android releases from caf using the linux kernel android for msm firefox os for msm qrd android before security patch level publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
399,950 | 27,263,071,008 | IssuesEvent | 2023-02-22 16:06:51 | nest-desktop/nest-desktop | https://api.github.com/repos/nest-desktop/nest-desktop | opened | Docs: Incorrect rendering in the Snap installation guide | bug documentation | **Describe the bug**
The setup for Snap (.../user/setup/index.html#snap-linux) in the docs is not rendered correctly.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the Snap page
2. See error
**Expected behavior**
The rendering should be correct.
**Screenshots**

**Desktop (please complete the following information):**
OS-independent (?)
**Smartphone (please complete the following information):**
OS-independent (?)
**Additional context**
n/a | 1.0 | Docs: Incorrect rendering in the Snap installation guide - **Describe the bug**
The setup for Snap (.../user/setup/index.html#snap-linux) in the docs is not rendered correctly.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the Snap page
2. See error
**Expected behavior**
The rendering should be correct.
**Screenshots**

**Desktop (please complete the following information):**
OS-independent (?)
**Smartphone (please complete the following information):**
OS-independent (?)
**Additional context**
n/a | non_priority | docs incorrect rendering in the snap installation guide describe the bug the setup for snap user setup index html snap linux in the docs is not rendered correctly to reproduce steps to reproduce the behavior visit the snap page see error expected behavior the rendering should be correct screenshots desktop please complete the following information os independent smartphone please complete the following information os independent additional context n a | 0 |
580,617 | 17,262,450,003 | IssuesEvent | 2021-07-22 09:30:01 | skni-kod/TournamentAppFrontend | https://api.github.com/repos/skni-kod/TournamentAppFrontend | closed | Tournament Info Map Bug | High priority | Naprawić wypisanie błędów w konsoli podczas przechodzenia na Tournament Info związanych z mapą - ewentualnie sprawdzić, czy błąd się nie pojawia, jeśli dane do mapy byłyby wczytywane dynamicznie z bazy (sam link).
Błąd na żółtym tle nie zawsze się pojawia - sprawdzić dlaczego.

| 1.0 | Tournament Info Map Bug - Naprawić wypisanie błędów w konsoli podczas przechodzenia na Tournament Info związanych z mapą - ewentualnie sprawdzić, czy błąd się nie pojawia, jeśli dane do mapy byłyby wczytywane dynamicznie z bazy (sam link).
Błąd na żółtym tle nie zawsze się pojawia - sprawdzić dlaczego.

| priority | tournament info map bug naprawić wypisanie błędów w konsoli podczas przechodzenia na tournament info związanych z mapą ewentualnie sprawdzić czy błąd się nie pojawia jeśli dane do mapy byłyby wczytywane dynamicznie z bazy sam link błąd na żółtym tle nie zawsze się pojawia sprawdzić dlaczego | 1 |
379,725 | 11,234,516,655 | IssuesEvent | 2020-01-09 05:29:47 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Move SelectorSpread priority logic to its Score plugin [Migration Phase 2] | kind/feature priority/important-soon sig/scheduling | /sig scheduling
/priority important-soon
SelectorSpread priority is now called from a plugin named DefaultPodTopologySpread. We need to move the logic to run in the plugin and remove SelectorSpread priority completely.
See #86399 as an example.
Part of #85822
This needs PostFilter Plugin to store the pod selector.
/assign @hprateek43 | 1.0 | Move SelectorSpread priority logic to its Score plugin [Migration Phase 2] - /sig scheduling
/priority important-soon
SelectorSpread priority is now called from a plugin named DefaultPodTopologySpread. We need to move the logic to run in the plugin and remove SelectorSpread priority completely.
See #86399 as an example.
Part of #85822
This needs PostFilter Plugin to store the pod selector.
/assign @hprateek43 | priority | move selectorspread priority logic to its score plugin sig scheduling priority important soon selectorspread priority is now called from a plugin named defaultpodtopologyspread we need to move the logic to run in the plugin and remove selectorspread priority completely see as an example part of this needs postfilter plugin to store the pod selector assign | 1 |
337,474 | 24,541,396,936 | IssuesEvent | 2022-10-12 04:19:00 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [Docs][API] Add CREATE TABLESPACE to YSQL API docs | area/documentation | ### Description
Add CREATE TABLESPACES to [YSQL API](https://docs.yugabyte.com/latest/api/ysql/the-sql-language/statements/) docs. | 1.0 | [Docs][API] Add CREATE TABLESPACE to YSQL API docs - ### Description
Add CREATE TABLESPACES to [YSQL API](https://docs.yugabyte.com/latest/api/ysql/the-sql-language/statements/) docs. | non_priority | add create tablespace to ysql api docs description add create tablespaces to docs | 0 |
557,142 | 16,502,015,240 | IssuesEvent | 2021-05-25 15:20:35 | NeuraLegion/nexploit-cli | https://api.github.com/repos/NeuraLegion/nexploit-cli | closed | Ignore remote scripts if local scripts are used | Priority: high Type: enhancement | Allow configure from a cloud of script settings, but if a local script is selected they should be ignored once the repeater is connected
| 1.0 | Ignore remote scripts if local scripts are used - Allow configure from a cloud of script settings, but if a local script is selected they should be ignored once the repeater is connected
| priority | ignore remote scripts if local scripts are used allow configure from a cloud of script settings but if a local script is selected they should be ignored once the repeater is connected | 1 |
346,826 | 24,887,271,592 | IssuesEvent | 2022-10-28 08:51:55 | brian-vb/ped | https://api.github.com/repos/brian-vb/ped | opened | Incorrect Instruction for Step 5 of UG (Quick Start) | severity.VeryLow type.DocumentationBug | 
The downloaded jar file is tp.jar
<!--session: 1666946577070-1af66f5c-5dec-424b-aed2-3d4e50698581-->
<!--Version: Web v3.4.4--> | 1.0 | Incorrect Instruction for Step 5 of UG (Quick Start) - 
The downloaded jar file is tp.jar
<!--session: 1666946577070-1af66f5c-5dec-424b-aed2-3d4e50698581-->
<!--Version: Web v3.4.4--> | non_priority | incorrect instruction for step of ug quick start the downloaded jar file is tp jar | 0 |
125,326 | 4,955,587,017 | IssuesEvent | 2016-12-01 20:51:27 | caitlynmayers/dukes | https://api.github.com/repos/caitlynmayers/dukes | opened | Shopping Cart: Create Account Alignment | Low Priority Style | Upon checking out, "create an account" is on a separate line than the check box
<img width="189" alt="screen shot 2016-12-01 at 3 47 42 pm" src="https://cloud.githubusercontent.com/assets/24302252/20812010/fac027ea-b7dd-11e6-9e92-bb4a97466f82.png">
| 1.0 | Shopping Cart: Create Account Alignment - Upon checking out, "create an account" is on a separate line than the check box
<img width="189" alt="screen shot 2016-12-01 at 3 47 42 pm" src="https://cloud.githubusercontent.com/assets/24302252/20812010/fac027ea-b7dd-11e6-9e92-bb4a97466f82.png">
| priority | shopping cart create account alignment upon checking out create an account is on a separate line than the check box img width alt screen shot at pm src | 1 |
17,492 | 2,615,145,540 | IssuesEvent | 2015-03-01 06:21:26 | chrsmith/html5rocks | https://api.github.com/repos/chrsmith/html5rocks | closed | Need hook script for site update | auto-migrated Maintenance Milestone-2 Priority-Low Type-Enhancement | ```
- one for updating/building the slides cache manifest
- one for .zip'ing the studio samples
First is done, but should be put into a .sh script on in place of running
appcfg.py update.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 2 Aug 2010 at 4:52 | 1.0 | Need hook script for site update - ```
- one for updating/building the slides cache manifest
- one for .zip'ing the studio samples
First is done, but should be put into a .sh script on in place of running
appcfg.py update.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 2 Aug 2010 at 4:52 | priority | need hook script for site update one for updating building the slides cache manifest one for zip ing the studio samples first is done but should be put into a sh script on in place of running appcfg py update original issue reported on code google com by ericbide com on aug at | 1 |
39,392 | 10,335,278,628 | IssuesEvent | 2019-09-03 10:11:53 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Protocol "https" not supported or disabled in libcurl | stat:awaiting response subtype: ubuntu/linux type:build/install | Hi Guys,
can anyone help me with the following issue
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
- TensorFlow installed from source.
- TensorFlow version: r1.11
- Python version: 2.7.12
- GCC/Compiler version (if compiling from source): GNU 5.4.0
**Describe the problem**
from top directory of tensorflow iam running the script as follows
./tensorflow/tools/ci_build/linux/cmake/run.sh
iam stuck with the following error
error: downloading 'https://mirror.bazel.build/ufpr.dl.sourceforge.net/project/giflib/giflib-5.1.4.tar.gz' failed
status_code: 1
status_string: "Unsupported protocol"
log:
--- LOG BEGIN ---
Protocol "https" not supported or disabled in libcurl
like above error there are number of files that are being not downloaded because of that iam getting build error.
this is my curl version iam using
curl -V
curl 7.66.0-DEV (x86_64-pc-linux-gnu) libcurl/7.66.0-DEV OpenSSL/1.0.2g zlib/1.2.8
Release-Date: [unreleased]
Protocols: dict file ftp ftps gopher http https imap imaps pop3 pop3s rtsp smb smbs smtp smtps telnet tftp
Features: AsynchDNS HTTPS-proxy IPv6 Largefile libz NTLM NTLM_WB SSL TLS-SRP UnixSockets
Thank you
Snathick
| 1.0 | Protocol "https" not supported or disabled in libcurl - Hi Guys,
can anyone help me with the following issue
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
- TensorFlow installed from source.
- TensorFlow version: r1.11
- Python version: 2.7.12
- GCC/Compiler version (if compiling from source): GNU 5.4.0
**Describe the problem**
from top directory of tensorflow iam running the script as follows
./tensorflow/tools/ci_build/linux/cmake/run.sh
iam stuck with the following error
error: downloading 'https://mirror.bazel.build/ufpr.dl.sourceforge.net/project/giflib/giflib-5.1.4.tar.gz' failed
status_code: 1
status_string: "Unsupported protocol"
log:
--- LOG BEGIN ---
Protocol "https" not supported or disabled in libcurl
like above error there are number of files that are being not downloaded because of that iam getting build error.
this is my curl version iam using
curl -V
curl 7.66.0-DEV (x86_64-pc-linux-gnu) libcurl/7.66.0-DEV OpenSSL/1.0.2g zlib/1.2.8
Release-Date: [unreleased]
Protocols: dict file ftp ftps gopher http https imap imaps pop3 pop3s rtsp smb smbs smtp smtps telnet tftp
Features: AsynchDNS HTTPS-proxy IPv6 Largefile libz NTLM NTLM_WB SSL TLS-SRP UnixSockets
Thank you
Snathick
| non_priority | protocol https not supported or disabled in libcurl hi guys can anyone help me with the following issue system information os platform and distribution e g linux ubuntu tensorflow installed from source tensorflow version python version gcc compiler version if compiling from source gnu describe the problem from top directory of tensorflow iam running the script as follows tensorflow tools ci build linux cmake run sh iam stuck with the following error error downloading failed status code status string unsupported protocol log log begin protocol https not supported or disabled in libcurl like above error there are number of files that are being not downloaded because of that iam getting build error this is my curl version iam using curl v curl dev pc linux gnu libcurl dev openssl zlib release date protocols dict file ftp ftps gopher http https imap imaps rtsp smb smbs smtp smtps telnet tftp features asynchdns https proxy largefile libz ntlm ntlm wb ssl tls srp unixsockets thank you snathick | 0 |
408,669 | 11,950,644,763 | IssuesEvent | 2020-04-03 15:32:04 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | closed | Disable SITS and rotate tests in the KubeCF pipeline | Priority: Critical Status: Accepted Type: CI | **Is your feature request related to a problem? Please describe.**
SITS and rotation tests are often failing, and they are gating tests that are passing
**Describe the solution you'd like**
Disable SITS and rotation tests until they prove to be reliable.
**Describe alternatives you've considered**
N/A
**Additional context**
Part of https://github.com/cloudfoundry-incubator/kubecf/issues/580
| 1.0 | Disable SITS and rotate tests in the KubeCF pipeline - **Is your feature request related to a problem? Please describe.**
SITS and rotation tests are often failing, and they are gating tests that are passing
**Describe the solution you'd like**
Disable SITS and rotation tests until they prove to be reliable.
**Describe alternatives you've considered**
N/A
**Additional context**
Part of https://github.com/cloudfoundry-incubator/kubecf/issues/580
| priority | disable sits and rotate tests in the kubecf pipeline is your feature request related to a problem please describe sits and rotation tests are often failing and they are gating tests that are passing describe the solution you d like disable sits and rotation tests until they prove to be reliable describe alternatives you ve considered n a additional context part of | 1 |
56,624 | 15,226,712,518 | IssuesEvent | 2021-02-18 09:15:33 | galasa-dev/projectmanagement | https://api.github.com/repos/galasa-dev/projectmanagement | closed | Galasa 3270 locking up during local runs | Manager: zOS 3270 defect | It appears that galasa seems to be locking up during local runs | 1.0 | Galasa 3270 locking up during local runs - It appears that galasa seems to be locking up during local runs | non_priority | galasa locking up during local runs it appears that galasa seems to be locking up during local runs | 0 |
26,511 | 4,733,369,539 | IssuesEvent | 2016-10-19 10:56:30 | sfepy/sfepy | https://api.github.com/repos/sfepy/sfepy | closed | "Latest snapshot" broken link | defect easy-to-fix | In the [website](http://sfepy.org/doc-devel/downloads.html) appears the following link
http://github.com/sfepy/sfepy/archives/master
But it leads to nowhere.
| 1.0 | "Latest snapshot" broken link - In the [website](http://sfepy.org/doc-devel/downloads.html) appears the following link
http://github.com/sfepy/sfepy/archives/master
But it leads to nowhere.
| non_priority | latest snapshot broken link in the appears the following link but it leads to nowhere | 0 |
322,847 | 9,829,323,639 | IssuesEvent | 2019-06-15 19:40:11 | marklogic/marklogic-data-hub | https://api.github.com/repos/marklogic/marklogic-data-hub | closed | Duplicate indices being created by QuickStart in entity-config/databases/final-database.json and staging-database.json | 2.0 3.0 4.0 BackLog better-errors bug priority:medium | [final-database.json.txt](https://github.com/marklogic-community/marklogic-data-hub/files/1471693/final-database.json.txt)
[staging-database.json.txt](https://github.com/marklogic-community/marklogic-data-hub/files/1471694/staging-database.json.txt)
#### The issue
Short description of the problem:
Duplicate indices being created by QuickStart in entity-config/databases/final-database.json and staging-database.json.
What behavior are you expecting?
I expect QuickStart to NOT create duplicate indices when the user creates 2 entities, and each having the same property (intentional).
#### Tech details
Which Operating System are you using?
* Windows 10
* AWS Linux (Linux ip-10-239-12-211 4.4.44-39.55.amzn1.x86_64 #1 SMP Mon Jan 30 18:15:53 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux)
Which version of MarkLogic are you using? Observed in
* 9.03
* 9.02
Which version of the Data Hub Framework are you using?
* DHF 2.0 Final
Is this a QuickStart UI bug? If so, which browser are you using?
* Duplicate indices observed through the use of QuickStart -- Chrome
* Error observed occasionally when using gradle headless after creating the entities and adding the same properties on both entities (intentially stating that the 2 entities have same property)
#### The devil is in the details...
See attachments, sample entity-config/databases/final-database.json and staging-database.json
Renamed with .txt suffixes so that these can be attached to the issue.
Error observed on occasion:
Error:
> Task :hubPreInstallCheck
com.marklogic.hub.PreInstallCheck@7c3c1335
> Task :mlDeployApp
Logging HTTP response body to assist with debugging: {"errorResponse":{"statusCode":"500",
"status":"Internal Server Error", "messageCode":"ADMIN-DUPLICATECONFIGITEM",
"message":"ADMIN-DUPLICATECONFIGITEM: (err:FOER0000) Database resource 'xxxx-data-
hub-FINAL' already contains 'Database 'range-element-index''"}}
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':mlDeployApp'.
> 500 Internal Server Error | 1.0 | Duplicate indices being created by QuickStart in entity-config/databases/final-database.json and staging-database.json - [final-database.json.txt](https://github.com/marklogic-community/marklogic-data-hub/files/1471693/final-database.json.txt)
[staging-database.json.txt](https://github.com/marklogic-community/marklogic-data-hub/files/1471694/staging-database.json.txt)
#### The issue
Short description of the problem:
Duplicate indices being created by QuickStart in entity-config/databases/final-database.json and staging-database.json.
What behavior are you expecting?
I expect QuickStart to NOT create duplicate indices when the user creates 2 entities, and each having the same property (intentional).
#### Tech details
Which Operating System are you using?
* Windows 10
* AWS Linux (Linux ip-10-239-12-211 4.4.44-39.55.amzn1.x86_64 #1 SMP Mon Jan 30 18:15:53 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux)
Which version of MarkLogic are you using? Observed in
* 9.03
* 9.02
Which version of the Data Hub Framework are you using?
* DHF 2.0 Final
Is this a QuickStart UI bug? If so, which browser are you using?
* Duplicate indices observed through the use of QuickStart -- Chrome
* Error observed occasionally when using gradle headless after creating the entities and adding the same properties on both entities (intentially stating that the 2 entities have same property)
#### The devil is in the details...
See attachments, sample entity-config/databases/final-database.json and staging-database.json
Renamed with .txt suffixes so that these can be attached to the issue.
Error observed on occasion:
Error:
> Task :hubPreInstallCheck
com.marklogic.hub.PreInstallCheck@7c3c1335
> Task :mlDeployApp
Logging HTTP response body to assist with debugging: {"errorResponse":{"statusCode":"500",
"status":"Internal Server Error", "messageCode":"ADMIN-DUPLICATECONFIGITEM",
"message":"ADMIN-DUPLICATECONFIGITEM: (err:FOER0000) Database resource 'xxxx-data-
hub-FINAL' already contains 'Database 'range-element-index''"}}
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':mlDeployApp'.
> 500 Internal Server Error | priority | duplicate indices being created by quickstart in entity config databases final database json and staging database json the issue short description of the problem duplicate indices being created by quickstart in entity config databases final database json and staging database json what behavior are you expecting i expect quickstart to not create duplicate indices when the user creates entities and each having the same property intentional tech details which operating system are you using windows aws linux linux ip smp mon jan utc gnu linux which version of marklogic are you using observed in which version of the data hub framework are you using dhf final is this a quickstart ui bug if so which browser are you using duplicate indices observed through the use of quickstart chrome error observed occasionally when using gradle headless after creating the entities and adding the same properties on both entities intentially stating that the entities have same property the devil is in the details see attachments sample entity config databases final database json and staging database json renamed with txt suffixes so that these can be attached to the issue error observed on occasion error task hubpreinstallcheck com marklogic hub preinstallcheck task mldeployapp logging http response body to assist with debugging errorresponse statuscode status internal server error messagecode admin duplicateconfigitem message admin duplicateconfigitem err database resource xxxx data hub final already contains database range element index failure build failed with an exception what went wrong execution failed for task mldeployapp internal server error | 1 |
649,907 | 21,329,970,371 | IssuesEvent | 2022-04-18 06:55:57 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | az1.qualtrics.com - design is broken | priority-important browser-fenix engine-gecko | <!-- @browser: Firefox Mobile 101.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:101.0) Gecko/101.0 Firefox/101.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/102794 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://az1.qualtrics.com/dg-service/account-setup?token=U2FsdGVkX19eeIZEkBiKWqPBDQxKZ7%2B%2FeK0sy%2FralpsMsw2WZfBDdBBNQ7ACC3QVZoppXM0dWjXZNv3cCEJgt6mimT2Dijen0AdZ%2BIJzdPv3SwIi1AzskYjbTDs6Q1RTIt6poDSgbcndhvpqEXgdDUsK9fg1z8OPuxWlPlgKkf4itxmo4nF1HpGj4PJUBFPlAY4i6UhxIZ6EtfvmHljpmW8VADKPdwYB%2FRXLp9Xk2%2BvXmr5ZyuTKXFgXz0cgO3YFGpexByNiTKM3qCsCx3QGKXWs3OsPMd2zQs1VYqdMcNtkIzXrWuXSykmj3gIzJHkcVz4s7hxpUAskA5Td8jvei4Aq6BEAXaV0VLMjYtTOYRDCQHfT5QtOJDt7D73RWAgifDbcMeN6o1YhZwsQtuB2bfFoiD2oCJ4zeAsSADGo4664KlrRffR65EpFM7wsxCkIv7M%2BDk3so7iJEykhM5%2FwqTi6Ri2%2F8bS5hK6GAJAa7ZiUtTP0V77NyANXFFG0NmKq2P8%2FJ90ARDJWIoyIv9NMd6nVVYsk86jQaoKeigg9aO%2FP0JZqxKIUYcrUoKsDcKErsT8YsFOLECdLPt%2BnctZUkbu4eJ7PRIGEmdfR6xfj284i5Pat4aDebsIbgbOhKqqptFW6So3cWySKrJ5wkBUkGzUsMoVdRUTCDzg1J6F35fIqHNddKUSRKTpntMI%2FJ97JtBBroP0s17SJ720l%2BL9wlKcrs2ca9Bnf%2FCIa5PZEVfJH5TuMM6zQLrkQ9tzDZmwOavNCgbKjmg70ijlmOOoyQ9V%2B%2FhuRJs8tyfZ%2BPUkrLGGw%2Fncm8EdDvS5DBh1cr6AKZgBGIIopU%2FJYVP%2BS6zGH8cJ4qe06ebqyQt%2FMvYXFqyoTGCS%2BlZFurhF2mmyCqWrtacd%2FMFPfbgBb8oG7smMWZbJKkCJeKiD4dVPuy7ecoEIQBhJyzeVz3EZB%2BAQbXOIPYtB%2FdBcpRGyew%2BdOzq8ImZdwnceWQdMXpYGDz7T%2BXVe5%2FPlnYxE3%2FqxBqP123grJssfZAduxc2ThdtinYCzWzOHZMxLg72nef7vvft2zCXgkDvoLIEHSE5D6XdDNIT5r&use_case=PERSONAL&info_case=
**Browser / Version**: Firefox Mobile 101.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
my kids are missing they tampered with a federal lawsuit I'm being followed they killed my mom and are hacking every device and If in October when this dumb cunt child molester shardae had not of sold me out or fucking a child molester williamrandy507@gmail.com or told me that it's my fault for standing up to racism or lying for another child molester bitch sold my accounts and passwords I want justice
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/4/b870d2f8-6a45-4f66-85d1-04a0c8472c4e.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220414092955</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/4/4632fe64-6aec-43f6-9cd5-cd89ac8a90ee)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | az1.qualtrics.com - design is broken - <!-- @browser: Firefox Mobile 101.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:101.0) Gecko/101.0 Firefox/101.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/102794 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://az1.qualtrics.com/dg-service/account-setup?token=U2FsdGVkX19eeIZEkBiKWqPBDQxKZ7%2B%2FeK0sy%2FralpsMsw2WZfBDdBBNQ7ACC3QVZoppXM0dWjXZNv3cCEJgt6mimT2Dijen0AdZ%2BIJzdPv3SwIi1AzskYjbTDs6Q1RTIt6poDSgbcndhvpqEXgdDUsK9fg1z8OPuxWlPlgKkf4itxmo4nF1HpGj4PJUBFPlAY4i6UhxIZ6EtfvmHljpmW8VADKPdwYB%2FRXLp9Xk2%2BvXmr5ZyuTKXFgXz0cgO3YFGpexByNiTKM3qCsCx3QGKXWs3OsPMd2zQs1VYqdMcNtkIzXrWuXSykmj3gIzJHkcVz4s7hxpUAskA5Td8jvei4Aq6BEAXaV0VLMjYtTOYRDCQHfT5QtOJDt7D73RWAgifDbcMeN6o1YhZwsQtuB2bfFoiD2oCJ4zeAsSADGo4664KlrRffR65EpFM7wsxCkIv7M%2BDk3so7iJEykhM5%2FwqTi6Ri2%2F8bS5hK6GAJAa7ZiUtTP0V77NyANXFFG0NmKq2P8%2FJ90ARDJWIoyIv9NMd6nVVYsk86jQaoKeigg9aO%2FP0JZqxKIUYcrUoKsDcKErsT8YsFOLECdLPt%2BnctZUkbu4eJ7PRIGEmdfR6xfj284i5Pat4aDebsIbgbOhKqqptFW6So3cWySKrJ5wkBUkGzUsMoVdRUTCDzg1J6F35fIqHNddKUSRKTpntMI%2FJ97JtBBroP0s17SJ720l%2BL9wlKcrs2ca9Bnf%2FCIa5PZEVfJH5TuMM6zQLrkQ9tzDZmwOavNCgbKjmg70ijlmOOoyQ9V%2B%2FhuRJs8tyfZ%2BPUkrLGGw%2Fncm8EdDvS5DBh1cr6AKZgBGIIopU%2FJYVP%2BS6zGH8cJ4qe06ebqyQt%2FMvYXFqyoTGCS%2BlZFurhF2mmyCqWrtacd%2FMFPfbgBb8oG7smMWZbJKkCJeKiD4dVPuy7ecoEIQBhJyzeVz3EZB%2BAQbXOIPYtB%2FdBcpRGyew%2BdOzq8ImZdwnceWQdMXpYGDz7T%2BXVe5%2FPlnYxE3%2FqxBqP123grJssfZAduxc2ThdtinYCzWzOHZMxLg72nef7vvft2zCXgkDvoLIEHSE5D6XdDNIT5r&use_case=PERSONAL&info_case=
**Browser / Version**: Firefox Mobile 101.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Design is broken
**Description**: Items are overlapped
**Steps to Reproduce**:
my kids are missing they tampered with a federal lawsuit I'm being followed they killed my mom and are hacking every device and If in October when this dumb cunt child molester shardae had not of sold me out or fucking a child molester williamrandy507@gmail.com or told me that it's my fault for standing up to racism or lying for another child molester bitch sold my accounts and passwords I want justice
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/4/b870d2f8-6a45-4f66-85d1-04a0c8472c4e.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220414092955</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/4/4632fe64-6aec-43f6-9cd5-cd89ac8a90ee)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | qualtrics com design is broken url browser version firefox mobile operating system android tested another browser yes internet explorer problem type design is broken description items are overlapped steps to reproduce my kids are missing they tampered with a federal lawsuit i m being followed they killed my mom and are hacking every device and if in october when this dumb cunt child molester shardae had not of sold me out or fucking a child molester gmail com or told me that it s my fault for standing up to racism or lying for another child molester bitch sold my accounts and passwords i want justice view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
206,098 | 7,108,570,063 | IssuesEvent | 2018-01-17 00:48:27 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [craftercms] The gradlew test stops to display the log messages send by test cases | bug priority: high | ### Expected behavior
The gradlew should not stops to displays log messages until the end of the execution.
### Actual behavior
After a time (6mins aprox) the gradlew stops to display the log messages send by each test cases and verbose of testNG. However, the automation continues executing in background.
And for the test process seems like it is stuck.
See the next image:

### Steps to reproduce the problem
* Go to your craftercms folder and run the next commands:
1. .\gradlew.bat selfupdate
2. .\gradlew.bat init
3. .\gradlew.bat update
4. .\gradlew.bat -P"authoring.studio.smtp.port=2525" build deploy
* Run the test using the next command:
1. .\gradlew.bat test
* Check and notice that the gradlew stops to display the log messages
Notes: I run it with maven and all the test cases log messages are displayed all the execution time
### Log/stack trace (use https://gist.github.com)
### Specs
#### Version
Crafter 3.0.3 and develop branch
#### OS
Windows Server 2012 R2, also for Windows 10 version
#### Browser
Chrome | 1.0 | [craftercms] The gradlew test stops to display the log messages send by test cases - ### Expected behavior
The gradlew should not stops to displays log messages until the end of the execution.
### Actual behavior
After a time (6mins aprox) the gradlew stops to display the log messages send by each test cases and verbose of testNG. However, the automation continues executing in background.
And for the test process seems like it is stuck.
See the next image:

### Steps to reproduce the problem
* Go to your craftercms folder and run the next commands:
1. .\gradlew.bat selfupdate
2. .\gradlew.bat init
3. .\gradlew.bat update
4. .\gradlew.bat -P"authoring.studio.smtp.port=2525" build deploy
* Run the test using the next command:
1. .\gradlew.bat test
* Check and notice that the gradlew stops to display the log messages
Notes: I run it with maven and all the test cases log messages are displayed all the execution time
### Log/stack trace (use https://gist.github.com)
### Specs
#### Version
Crafter 3.0.3 and develop branch
#### OS
Windows Server 2012 R2, also for Windows 10 version
#### Browser
Chrome | priority | the gradlew test stops to display the log messages send by test cases expected behavior the gradlew should not stops to displays log messages until the end of the execution actual behavior after a time aprox the gradlew stops to display the log messages send by each test cases and verbose of testng however the automation continues executing in background and for the test process seems like it is stuck see the next image steps to reproduce the problem go to your craftercms folder and run the next commands gradlew bat selfupdate gradlew bat init gradlew bat update gradlew bat p authoring studio smtp port build deploy run the test using the next command gradlew bat test check and notice that the gradlew stops to display the log messages notes i run it with maven and all the test cases log messages are displayed all the execution time log stack trace use specs version crafter and develop branch os windows server also for windows version browser chrome | 1 |
38,441 | 10,196,547,258 | IssuesEvent | 2019-08-12 21:02:43 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Lock file missing entries when RuntimeIdentifier is specified for netcoreapp 3.0 projects | Area:RepeatableBuild Area:Restore ClosedAs:Duplicate Style:PackageReference Type:Bug | ## Details about Problem
NuGet product used: dotnet.exe
dotnet.exe --version: 3.0.100-preview7-012821
Worked before? If so, with which NuGet version: yes. Seems to work even with current sdk if targeting netcoreapp2.0
## Detailed repro steps so we can see the same problem
### Automated repro
Download [repro_missingRuntimePackage.zip](https://github.com/NuGet/Home/files/3488060/repro_missingRuntimePackage.zip) and run repro.bat for an automated repro.
### Manual steps
1. Create a basic .csproj targeting netcoreapp3.0
2. Add `<RuntimeIdentifier>win-x64</RuntimeIdentifier>`
3. `dotnet restore --use-lock-file`
4. Notice that the generated lock file has no packages in it even though the build actually uses several (e.g. microsoft.netcore.app.runtime.win-x64)
This is a regression from netcoreapp2.0, which listed all packages as expected.
Tagging @rrelyea and @nkolev92 as FYI.
| 1.0 | Lock file missing entries when RuntimeIdentifier is specified for netcoreapp 3.0 projects - ## Details about Problem
NuGet product used: dotnet.exe
dotnet.exe --version: 3.0.100-preview7-012821
Worked before? If so, with which NuGet version: yes. Seems to work even with current sdk if targeting netcoreapp2.0
## Detailed repro steps so we can see the same problem
### Automated repro
Download [repro_missingRuntimePackage.zip](https://github.com/NuGet/Home/files/3488060/repro_missingRuntimePackage.zip) and run repro.bat for an automated repro.
### Manual steps
1. Create a basic .csproj targeting netcoreapp3.0
2. Add `<RuntimeIdentifier>win-x64</RuntimeIdentifier>`
3. `dotnet restore --use-lock-file`
4. Notice that the generated lock file has no packages in it even though the build actually uses several (e.g. microsoft.netcore.app.runtime.win-x64)
This is a regression from netcoreapp2.0, which listed all packages as expected.
Tagging @rrelyea and @nkolev92 as FYI.
| non_priority | lock file missing entries when runtimeidentifier is specified for netcoreapp projects details about problem nuget product used dotnet exe dotnet exe version worked before if so with which nuget version yes seems to work even with current sdk if targeting detailed repro steps so we can see the same problem automated repro download and run repro bat for an automated repro manual steps create a basic csproj targeting add win dotnet restore use lock file notice that the generated lock file has no packages in it even though the build actually uses several e g microsoft netcore app runtime win this is a regression from which listed all packages as expected tagging rrelyea and as fyi | 0 |
156,532 | 24,624,567,437 | IssuesEvent | 2022-10-16 10:52:22 | dotnet/efcore | https://api.github.com/repos/dotnet/efcore | closed | Dictinct not working for custom select | closed-by-design | Dictinct not add in SQL
For query:
```
var queryable = this.dbSet.Where(
x =>
x.Translation.StartsWith(model.Word) &&
(x.Localization == model.From || x.Localization == model.Dest)).Distinct()
.Select(x => new DictionaryWordCompleterApiModel
{
Word = x.Translation,
Localization = x.Localization
}).Take(10).Distinct();
```
I have DISTINCT in SQL:
```
SELECT TOP(10) [t].[Translation], [t].[Localization]
FROM (
SELECT DISTINCT [x0].*
FROM [GameDictionaryWords] AS [x0]
WHERE ([x0].[Translation] LIKE N'guide' + N'%' AND (CHARINDEX(N'guide', [x0].[Translation]) = 1)) AND [x0].[Localization] IN (N'en', N'es')
) AS [t]
```
BUT for query
```
var queryable = this.dbSet.Where(
x =>
x.Translation.StartsWith(model.Word) &&
(x.Localization == model.From || x.Localization == model.Dest))
.Select(x => new DictionaryWordCompleterApiModel
{
Word = x.Translation,
Localization = x.Localization
}).Take(10).Distinct();
```
not DISTINCT in SQL:
```
SELECT TOP(10) [x].[Translation], [x].[Localization]
FROM [GameDictionaryWords] AS [x]
WHERE ([x].[Translation] LIKE N'guid' + N'%' AND (CHARINDEX(N'guid', [x].[Translation]) = 1)) AND [x].[Localization] IN (N'en', N'es')
```
Why not add Distinct?
### Further technical details
EF Core version: 1.1.0
Database Provider: Microsoft.EntityFrameworkCore.SqlServer
Operating system: Windows 10
IDE: Visual Studio 2015
| 1.0 | Dictinct not working for custom select - Dictinct not add in SQL
For query:
```
var queryable = this.dbSet.Where(
x =>
x.Translation.StartsWith(model.Word) &&
(x.Localization == model.From || x.Localization == model.Dest)).Distinct()
.Select(x => new DictionaryWordCompleterApiModel
{
Word = x.Translation,
Localization = x.Localization
}).Take(10).Distinct();
```
I have DISTINCT in SQL:
```
SELECT TOP(10) [t].[Translation], [t].[Localization]
FROM (
SELECT DISTINCT [x0].*
FROM [GameDictionaryWords] AS [x0]
WHERE ([x0].[Translation] LIKE N'guide' + N'%' AND (CHARINDEX(N'guide', [x0].[Translation]) = 1)) AND [x0].[Localization] IN (N'en', N'es')
) AS [t]
```
BUT for query
```
var queryable = this.dbSet.Where(
x =>
x.Translation.StartsWith(model.Word) &&
(x.Localization == model.From || x.Localization == model.Dest))
.Select(x => new DictionaryWordCompleterApiModel
{
Word = x.Translation,
Localization = x.Localization
}).Take(10).Distinct();
```
not DISTINCT in SQL:
```
SELECT TOP(10) [x].[Translation], [x].[Localization]
FROM [GameDictionaryWords] AS [x]
WHERE ([x].[Translation] LIKE N'guid' + N'%' AND (CHARINDEX(N'guid', [x].[Translation]) = 1)) AND [x].[Localization] IN (N'en', N'es')
```
Why not add Distinct?
### Further technical details
EF Core version: 1.1.0
Database Provider: Microsoft.EntityFrameworkCore.SqlServer
Operating system: Windows 10
IDE: Visual Studio 2015
| non_priority | dictinct not working for custom select dictinct not add in sql for query var queryable this dbset where x x translation startswith model word x localization model from x localization model dest distinct select x new dictionarywordcompleterapimodel word x translation localization x localization take distinct i have distinct in sql select top from select distinct from as where like n guide n and charindex n guide and in n en n es as but for query var queryable this dbset where x x translation startswith model word x localization model from x localization model dest select x new dictionarywordcompleterapimodel word x translation localization x localization take distinct not distinct in sql select top from as where like n guid n and charindex n guid and in n en n es why not add distinct further technical details ef core version database provider microsoft entityframeworkcore sqlserver operating system windows ide visual studio | 0 |
202,781 | 7,054,973,431 | IssuesEvent | 2018-01-04 04:59:10 | Marri/glowfic | https://api.github.com/repos/Marri/glowfic | opened | Move from gmail address to glowfic.com addresses for email | 3. medium priority 8. medium type: enhancement v. planned | We have now verified glowfic.com with Amazon SES, and should send mail from our domain instead of the Gmail account. | 1.0 | Move from gmail address to glowfic.com addresses for email - We have now verified glowfic.com with Amazon SES, and should send mail from our domain instead of the Gmail account. | priority | move from gmail address to glowfic com addresses for email we have now verified glowfic com with amazon ses and should send mail from our domain instead of the gmail account | 1 |
706,316 | 24,264,732,117 | IssuesEvent | 2022-09-28 04:34:55 | hackforla/expunge-assist | https://api.github.com/repos/hackforla/expunge-assist | opened | Sweeping layout change desktop top content padding | role: development priority: high | During the dev + design meeting when dev displayed the large distance between the progress bar and the start of the content on the final page of the letter generator, the design team agreed there was too much white space and the distance should be shortened to 56px. For consistency, we would like all pages on the desktop to have top padding of 56px for content.
- [ ] Adjust the top padding of the content boxes on all pages to 56px. to allow slightly more spacing from the progress bar. | 1.0 | Sweeping layout change desktop top content padding - During the dev + design meeting when dev displayed the large distance between the progress bar and the start of the content on the final page of the letter generator, the design team agreed there was too much white space and the distance should be shortened to 56px. For consistency, we would like all pages on the desktop to have top padding of 56px for content.
- [ ] Adjust the top padding of the content boxes on all pages to 56px. to allow slightly more spacing from the progress bar. | priority | sweeping layout change desktop top content padding during the dev design meeting when dev displayed the large distance between the progress bar and the start of the content on the final page of the letter generator the design team agreed there was too much white space and the distance should be shortened to for consistency we would like all pages on the desktop to have top padding of for content adjust the top padding of the content boxes on all pages to to allow slightly more spacing from the progress bar | 1 |
558,677 | 16,540,347,719 | IssuesEvent | 2021-05-27 16:02:27 | CredentialEngine/CredentialRegistry | https://api.github.com/repos/CredentialEngine/CredentialRegistry | closed | Getting a count of total results in SPARQL is extremely slow | Blocker High Priority | What we're seeing currently:
- Queries like "all of the credentials with a given SOC code" take about 45 seconds
- I took the query apart into its two pieces, "all credentials" and "things with a given SOC code", and:
- "all credentials" takes about a full minute(!) to execute
- "things with a given SOC code" takes about 18 seconds
That first one is really surprising, and a problem. Even if I:
- Remove the only traversal in the query (the one that fetches the payload for each match)
- Remove the part of the query that gets a count of all matches
- Add a global limit of 5 items to the very end of the query, outside all of its scopes
It still takes over 50 seconds.
If I add a limit of 5 to the inner query, then I can get the results back in ~2-4 seconds, but that also breaks the total results count (as it will never show a total greater than 5 since SPARQL stops counting when it hits that inner limit).
In previous versions of how this stuff gets built, I had tried running two queries (one for the data and one for the count, but otherwise identical) as separate but simultaneous requests. It seemed like one slowed the other down so there was no meaningful difference in the time it took to do that (but that was in an earlier SPARQL engine version). It seems the problem is in the way the total number of results is fetched.
Out of curiosity, I tried running the same query but just returning the count (no results) - it took almost a minute and a half.
The queries in question:
"Get all of the credentials for a given SOC code":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload (SUM(COALESCE(?relevance_points_947cd59921884436aef740734d208937, 0)) AS ?relevance_score) WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { VALUES ?360084717ea14cb1bfecf727052ae9c4 { ceterms:ApprenticeshipCertificate ceterms:AssociateDegree ceterms:BachelorDegree ceterms:Badge ceterms:Certificate ceterms:CertificateOfCompletion ceterms:Certification ceterms:Degree ceterms:DigitalBadge ceterms:Diploma ceterms:DoctoralDegree ceterms:GeneralEducationDevelopment ceterms:JourneymanCertificate ceterms:License ceterms:MasterCertificate ceterms:MasterDegree ceterms:MicroCredential ceterms:OpenBadge ceterms:ProfessionalDoctorate ceterms:QualityAssuranceCredential ceterms:ResearchDoctorate ceterms:SecondarySchoolDiploma } ?id a ?360084717ea14cb1bfecf727052ae9c4 . } { ?id ( (ceterms:occupationType/^credreg:__graph?) | (credreg:__graph?/^ceterms:occupationType) ) ?0911544d1a2647d482503b4fc3940725 . { ?0911544d1a2647d482503b4fc3940725 ( ceterms:codedNotation__tokenData ) ?tokens_e65b9c7f82554b03ac58637e89c62216 . { { ?tokens_e65b9c7f82554b03ac58637e89c62216 ( credreg:__tokenFullNormalized ) ?normalized_ee32d46bee8b4e79af50bab0a1b92ae2 . BIND(IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13|20'),IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13 20'), 1250, IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, ''), 125, IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13'), 4, 0) + IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '20'), 4, 0))), 0) AS ?relevance_points_947cd59921884436aef740734d208937) FILTER(?relevance_points_947cd59921884436aef740734d208937 > 0) } } } } } ?id credreg:__payload ?searchResultPayload . } GROUP BY ?id ?searchResultPayload ?relevance_score } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?relevance_score) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?relevance_score)
```
"Get all of the credentials":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { VALUES ?c9fa2af5263b4d019c53f582f70e6d30 { ceterms:ApprenticeshipCertificate ceterms:AssociateDegree ceterms:BachelorDegree ceterms:Badge ceterms:Certificate ceterms:CertificateOfCompletion ceterms:Certification ceterms:Degree ceterms:DigitalBadge ceterms:Diploma ceterms:DoctoralDegree ceterms:GeneralEducationDevelopment ceterms:JourneymanCertificate ceterms:License ceterms:MasterCertificate ceterms:MasterDegree ceterms:MicroCredential ceterms:OpenBadge ceterms:ProfessionalDoctorate ceterms:QualityAssuranceCredential ceterms:ResearchDoctorate ceterms:SecondarySchoolDiploma } ?id a ?c9fa2af5263b4d019c53f582f70e6d30 . } } ?id credreg:__payload ?searchResultPayload . } } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?id) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?id)
```
"Get all of the things with a given SOC code":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload (SUM(COALESCE(?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63, 0)) AS ?relevance_score) WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { ?id ( (ceterms:occupationType/^credreg:__graph?) | (credreg:__graph?/^ceterms:occupationType) ) ?b676cdb35aeb4f8cb7dd25ab1305199f . { ?b676cdb35aeb4f8cb7dd25ab1305199f ( ceterms:codedNotation__tokenData ) ?tokens_a5f00c8fde3847b3be16352bc05851d2 . { { ?tokens_a5f00c8fde3847b3be16352bc05851d2 ( credreg:__tokenFullNormalized ) ?normalized_91c056153e864b1e963de812893e362b . BIND(IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13|20'),IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13 20'), 1250, IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, ''), 125, IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13'), 4, 0) + IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '20'), 4, 0))), 0) AS ?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63) FILTER(?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63 > 0) } } } } } ?id credreg:__payload ?searchResultPayload . } GROUP BY ?id ?searchResultPayload ?relevance_score } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?relevance_score) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?relevance_score)
```
By far, the slowest part of it seems to be the COUNT(), even when there are no traversals. That seems odd to me. | 1.0 | Getting a count of total results in SPARQL is extremely slow - What we're seeing currently:
- Queries like "all of the credentials with a given SOC code" take about 45 seconds
- I took the query apart into its two pieces, "all credentials" and "things with a given SOC code", and:
- "all credentials" takes about a full minute(!) to execute
- "things with a given SOC code" takes about 18 seconds
That first one is really surprising, and a problem. Even if I:
- Remove the only traversal in the query (the one that fetches the payload for each match)
- Remove the part of the query that gets a count of all matches
- Add a global limit of 5 items to the very end of the query, outside all of its scopes
It still takes over 50 seconds.
If I add a limit of 5 to the inner query, then I can get the results back in ~2-4 seconds, but that also breaks the total results count (as it will never show a total greater than 5 since SPARQL stops counting when it hits that inner limit).
In previous versions of how this stuff gets built, I had tried running two queries (one for the data and one for the count, but otherwise identical) as separate but simultaneous requests. It seemed like one slowed the other down so there was no meaningful difference in the time it took to do that (but that was in an earlier SPARQL engine version). It seems the problem is in the way the total number of results is fetched.
Out of curiosity, I tried running the same query but just returning the count (no results) - it took almost a minute and a half.
The queries in question:
"Get all of the credentials for a given SOC code":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload (SUM(COALESCE(?relevance_points_947cd59921884436aef740734d208937, 0)) AS ?relevance_score) WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { VALUES ?360084717ea14cb1bfecf727052ae9c4 { ceterms:ApprenticeshipCertificate ceterms:AssociateDegree ceterms:BachelorDegree ceterms:Badge ceterms:Certificate ceterms:CertificateOfCompletion ceterms:Certification ceterms:Degree ceterms:DigitalBadge ceterms:Diploma ceterms:DoctoralDegree ceterms:GeneralEducationDevelopment ceterms:JourneymanCertificate ceterms:License ceterms:MasterCertificate ceterms:MasterDegree ceterms:MicroCredential ceterms:OpenBadge ceterms:ProfessionalDoctorate ceterms:QualityAssuranceCredential ceterms:ResearchDoctorate ceterms:SecondarySchoolDiploma } ?id a ?360084717ea14cb1bfecf727052ae9c4 . } { ?id ( (ceterms:occupationType/^credreg:__graph?) | (credreg:__graph?/^ceterms:occupationType) ) ?0911544d1a2647d482503b4fc3940725 . { ?0911544d1a2647d482503b4fc3940725 ( ceterms:codedNotation__tokenData ) ?tokens_e65b9c7f82554b03ac58637e89c62216 . { { ?tokens_e65b9c7f82554b03ac58637e89c62216 ( credreg:__tokenFullNormalized ) ?normalized_ee32d46bee8b4e79af50bab0a1b92ae2 . BIND(IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13|20'),IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13 20'), 1250, IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, ''), 125, IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '13'), 4, 0) + IF(REGEX(?normalized_ee32d46bee8b4e79af50bab0a1b92ae2, '20'), 4, 0))), 0) AS ?relevance_points_947cd59921884436aef740734d208937) FILTER(?relevance_points_947cd59921884436aef740734d208937 > 0) } } } } } ?id credreg:__payload ?searchResultPayload . } GROUP BY ?id ?searchResultPayload ?relevance_score } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?relevance_score) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?relevance_score)
```
"Get all of the credentials":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { VALUES ?c9fa2af5263b4d019c53f582f70e6d30 { ceterms:ApprenticeshipCertificate ceterms:AssociateDegree ceterms:BachelorDegree ceterms:Badge ceterms:Certificate ceterms:CertificateOfCompletion ceterms:Certification ceterms:Degree ceterms:DigitalBadge ceterms:Diploma ceterms:DoctoralDegree ceterms:GeneralEducationDevelopment ceterms:JourneymanCertificate ceterms:License ceterms:MasterCertificate ceterms:MasterDegree ceterms:MicroCredential ceterms:OpenBadge ceterms:ProfessionalDoctorate ceterms:QualityAssuranceCredential ceterms:ResearchDoctorate ceterms:SecondarySchoolDiploma } ?id a ?c9fa2af5263b4d019c53f582f70e6d30 . } } ?id credreg:__payload ?searchResultPayload . } } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?id) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?id)
```
"Get all of the things with a given SOC code":
```
PREFIX credreg: <https://credreg.net/> PREFIX ceterms: <https://purl.org/ctdl/terms/> SELECT ?totalResults ?id ?searchResultPayload ?relevance_score WITH { SELECT DISTINCT ?id ?searchResultPayload (SUM(COALESCE(?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63, 0)) AS ?relevance_score) WHERE { <http://aws.amazon.com/neptune/vocab/v01/QueryHints#Query> <http://aws.amazon.com/neptune/vocab/v01/QueryHints#joinOrder> 'Ordered' . ?id ceterms:ctid ?anyValue . { { ?id ( (ceterms:occupationType/^credreg:__graph?) | (credreg:__graph?/^ceterms:occupationType) ) ?b676cdb35aeb4f8cb7dd25ab1305199f . { ?b676cdb35aeb4f8cb7dd25ab1305199f ( ceterms:codedNotation__tokenData ) ?tokens_a5f00c8fde3847b3be16352bc05851d2 . { { ?tokens_a5f00c8fde3847b3be16352bc05851d2 ( credreg:__tokenFullNormalized ) ?normalized_91c056153e864b1e963de812893e362b . BIND(IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13|20'),IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13 20'), 1250, IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, ''), 125, IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '13'), 4, 0) + IF(REGEX(?normalized_91c056153e864b1e963de812893e362b, '20'), 4, 0))), 0) AS ?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63) FILTER(?relevance_points_51f0d3ec663b4d67a1cf28c679b88b63 > 0) } } } } } ?id credreg:__payload ?searchResultPayload . } GROUP BY ?id ?searchResultPayload ?relevance_score } AS %mainQuery WHERE { { SELECT (COUNT(DISTINCT ?id) AS ?totalResults) WHERE { INCLUDE %mainQuery } } UNION { SELECT ?id ?searchResultPayload ?relevance_score ?recordDate WHERE { INCLUDE %mainQuery } ORDER BY DESC(?relevance_score) OFFSET 0 LIMIT 5 } } ORDER BY DESC(?relevance_score)
```
By far, the slowest part of it seems to be the COUNT(), even when there are no traversals. That seems odd to me. | priority | getting a count of total results in sparql is extremely slow what we re seeing currently queries like all of the credentials with a given soc code take about seconds i took the query apart into its two pieces all credentials and things with a given soc code and all credentials takes about a full minute to execute things with a given soc code takes about seconds that first one is really surprising and a problem even if i remove the only traversal in the query the one that fetches the payload for each match remove the part of the query that gets a count of all matches add a global limit of items to the very end of the query outside all of its scopes it still takes over seconds if i add a limit of to the inner query then i can get the results back in seconds but that also breaks the total results count as it will never show a total greater than since sparql stops counting when it hits that inner limit in previous versions of how this stuff gets built i had tried running two queries one for the data and one for the count but otherwise identical as separate but simultaneous requests it seemed like one slowed the other down so there was no meaningful difference in the time it took to do that but that was in an earlier sparql engine version it seems the problem is in the way the total number of results is fetched out of curiosity i tried running the same query but just returning the count no results it took almost a minute and a half the queries in question get all of the credentials for a given soc code prefix credreg id credreg payload searchresultpayload group by id searchresultpayload relevance score as mainquery where select count distinct id as totalresults where include mainquery union select id searchresultpayload relevance score recorddate where include mainquery order by desc relevance score offset limit order by desc relevance score get all of the credentials prefix credreg prefix ceterms select totalresults id searchresultpayload relevance score with select distinct id searchresultpayload where ordered id ceterms ctid anyvalue values ceterms apprenticeshipcertificate ceterms associatedegree ceterms bachelordegree ceterms badge ceterms certificate ceterms certificateofcompletion ceterms certification ceterms degree ceterms digitalbadge ceterms diploma ceterms doctoraldegree ceterms generaleducationdevelopment ceterms journeymancertificate ceterms license ceterms mastercertificate ceterms masterdegree ceterms microcredential ceterms openbadge ceterms professionaldoctorate ceterms qualityassurancecredential ceterms researchdoctorate ceterms secondaryschooldiploma id a id credreg payload searchresultpayload as mainquery where select count distinct id as totalresults where include mainquery union select id searchresultpayload relevance score recorddate where include mainquery order by desc id offset limit order by desc id get all of the things with a given soc code prefix credreg id credreg payload searchresultpayload group by id searchresultpayload relevance score as mainquery where select count distinct id as totalresults where include mainquery union select id searchresultpayload relevance score recorddate where include mainquery order by desc relevance score offset limit order by desc relevance score by far the slowest part of it seems to be the count even when there are no traversals that seems odd to me | 1 |
168,267 | 14,144,494,291 | IssuesEvent | 2020-11-10 16:30:55 | dimosp/CineFriends | https://api.github.com/repos/dimosp/CineFriends | opened | Add the new post API endpoint functionality to the Documentation | documentation | Update Swagger and POSTMAN collection.
We have to include:
// photo
router.get('/posts/photo/:postId', photo);
// post routes
router.put('/posts/:postId', requireSignin, isPoster, updatePost);
router.delete('/posts/:postId', requireSignin, isPoster, deletePost); | 1.0 | Add the new post API endpoint functionality to the Documentation - Update Swagger and POSTMAN collection.
We have to include:
// photo
router.get('/posts/photo/:postId', photo);
// post routes
router.put('/posts/:postId', requireSignin, isPoster, updatePost);
router.delete('/posts/:postId', requireSignin, isPoster, deletePost); | non_priority | add the new post api endpoint functionality to the documentation update swagger and postman collection we have to include photo router get posts photo postid photo post routes router put posts postid requiresignin isposter updatepost router delete posts postid requiresignin isposter deletepost | 0 |
47,987 | 13,300,530,352 | IssuesEvent | 2020-08-25 11:31:05 | rammatzkvosky/biojava | https://api.github.com/repos/rammatzkvosky/biojava | opened | CVE-2019-20330 (High) detected in jackson-databind-2.9.9.3.jar | security vulnerability | ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/biojava/biojava-aa-prop/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- mmtf-serialization-1.0.9.jar (Root Library)
- jackson-dataformat-msgpack-0.8.18.jar
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/biojava/commit/ca34f598328cc90dbd6fc838853a5161ff47ecb9">ca34f598328cc90dbd6fc838853a5161ff47ecb9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","isTransitiveDependency":true,"dependencyTree":"org.rcsb:mmtf-serialization:1.0.9;org.msgpack:jackson-dataformat-msgpack:0.8.18;com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-20330 (High) detected in jackson-databind-2.9.9.3.jar - ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/biojava/biojava-aa-prop/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9.3/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- mmtf-serialization-1.0.9.jar (Root Library)
- jackson-dataformat-msgpack-0.8.18.jar
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/biojava/commit/ca34f598328cc90dbd6fc838853a5161ff47ecb9">ca34f598328cc90dbd6fc838853a5161ff47ecb9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","isTransitiveDependency":true,"dependencyTree":"org.rcsb:mmtf-serialization:1.0.9;org.msgpack:jackson-dataformat-msgpack:0.8.18;com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm biojava biojava aa prop pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy mmtf serialization jar root library jackson dataformat msgpack jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before lacks certain net sf ehcache blocking publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before lacks certain net sf ehcache blocking vulnerabilityurl | 0 |
75,669 | 7,480,280,267 | IssuesEvent | 2018-04-04 16:53:52 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Deletion of GKE cluster results in repeated errors logs in server relating to cluster not found. | kind/bug status/resolved status/to-test version/2.0 | Rancher versions: Build from master
Steps to reproduce the problem:
Created a GKE cluster.
Added 1 workload.
Deleted this cluster.
This results in repeated errors logs in server relating to cluster not found.
```
E0404 02:04:48.149453 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.158171 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.158286 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.380356 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-controllermanager [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.383907 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-deploment-event [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.385438 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-etcd [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.388824 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-node-mem [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.398725 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-scheduler [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.517202 1 generic_controller.go:204] ProjectAlertController project-b5dq2/projectalert-workload [project-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.107730 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.RoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/rolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.119945 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.PodSecurityPolicy: Get https://104.154.103.101/apis/extensions/v1beta1/podsecuritypolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130221 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Namespace: Get https://104.154.103.101/api/v1/namespaces?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130337 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRole: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterroles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130467 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ServiceAccount: Get https://104.154.103.101/api/v1/serviceaccounts?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.133910 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Pod: Get https://104.154.103.101/api/v1/pods?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.134030 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.DaemonSet: Get https://104.154.103.101/apis/apps/v1beta2/daemonsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.134130 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ReplicationController: Get https://104.154.103.101/api/v1/replicationcontrollers?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.148994 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.CronJob: Get https://104.154.103.101/apis/batch/v1beta1/cronjobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149127 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.Deployment: Get https://104.154.103.101/apis/apps/v1beta2/deployments?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149241 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Job: Get https://104.154.103.101/apis/batch/v1/jobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149351 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Endpoints: Get https://104.154.103.101/api/v1/endpoints?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149436 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.Ingress: Get https://104.154.103.101/apis/extensions/v1beta1/ingresses?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149550 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.NetworkPolicy: Get https://104.154.103.101/apis/networking.k8s.io/v1/networkpolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149636 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Event: Get https://104.154.103.101/api/v1/events?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149721 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.StatefulSet: Get https://104.154.103.101/apis/apps/v1beta2/statefulsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149807 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.ReplicaSet: Get https://104.154.103.101/apis/apps/v1beta2/replicasets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149886 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Secret: Get https://104.154.103.101/api/v1/secrets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149968 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Role: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/roles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.154805 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.158486 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.165613 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.108145 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.RoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/rolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.120265 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.PodSecurityPolicy: Get https://104.154.103.101/apis/extensions/v1beta1/podsecuritypolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.130455 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Namespace: Get https://104.154.103.101/api/v1/namespaces?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.134251 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRole: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterroles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.149426 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ServiceAccount: Get https://104.154.103.101/api/v1/serviceaccounts?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.155033 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Pod: Get https://104.154.103.101/api/v1/pods?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.158272 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.DaemonSet: Get https://104.154.103.101/apis/apps/v1beta2/daemonsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.159290 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ReplicationController: Get https://104.154.103.101/api/v1/replicationcontrollers?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.165706 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.CronJob: Get https://104.154.103.101/apis/batch/v1beta1/cronjobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.166768 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.Deployment: Get https://104.154.103.101/apis/apps/v1beta2/deployments?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.167814 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Job: Get https://104.154.103.101/apis/batch/v1/jobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.168880 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Endpoints: Get https://104.154.103.101/api/v1/endpoints?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.170327 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.Ingress: Get https://104.154.103.101/apis/extensions/v1beta1/ingresses?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.171910 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.NetworkPolicy: Get https://104.154.103.101/apis/networking.k8s.io/v1/networkpolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.179251 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Event: Get https://104.154.103.101/api/v1/events?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.180306 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.StatefulSet: Get https://104.154.103.101/apis/apps/v1beta2/statefulsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.181642 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.ReplicaSet: Get https://104.154.103.101/apis/apps/v1beta2/replicasets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.182628 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Secret: Get https://104.154.103.101/api/v1/secrets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.183720 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Role: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/roles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.185095 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.186151 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.187425 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
``` | 1.0 | Deletion of GKE cluster results in repeated errors logs in server relating to cluster not found. - Rancher versions: Build from master
Steps to reproduce the problem:
Created a GKE cluster.
Added 1 workload.
Deleted this cluster.
This results in repeated errors logs in server relating to cluster not found.
```
E0404 02:04:48.149453 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.158171 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.158286 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.380356 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-controllermanager [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.383907 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-deploment-event [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.385438 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-etcd [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.388824 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-node-mem [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.398725 1 generic_controller.go:204] ClusterAlertController cluster-n9gwr/clusteralert-scheduler [cluster-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:48.517202 1 generic_controller.go:204] ProjectAlertController project-b5dq2/projectalert-workload [project-alert-deployer] failed with : Delete https://104.154.103.101/api/v1/namespaces/cattle-alerting: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.107730 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.RoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/rolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.119945 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.PodSecurityPolicy: Get https://104.154.103.101/apis/extensions/v1beta1/podsecuritypolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130221 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Namespace: Get https://104.154.103.101/api/v1/namespaces?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130337 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRole: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterroles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.130467 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ServiceAccount: Get https://104.154.103.101/api/v1/serviceaccounts?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.133910 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Pod: Get https://104.154.103.101/api/v1/pods?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.134030 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.DaemonSet: Get https://104.154.103.101/apis/apps/v1beta2/daemonsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.134130 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ReplicationController: Get https://104.154.103.101/api/v1/replicationcontrollers?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.148994 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.CronJob: Get https://104.154.103.101/apis/batch/v1beta1/cronjobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149127 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.Deployment: Get https://104.154.103.101/apis/apps/v1beta2/deployments?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149241 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Job: Get https://104.154.103.101/apis/batch/v1/jobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149351 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Endpoints: Get https://104.154.103.101/api/v1/endpoints?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149436 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.Ingress: Get https://104.154.103.101/apis/extensions/v1beta1/ingresses?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149550 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.NetworkPolicy: Get https://104.154.103.101/apis/networking.k8s.io/v1/networkpolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149636 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Event: Get https://104.154.103.101/api/v1/events?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149721 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.StatefulSet: Get https://104.154.103.101/apis/apps/v1beta2/statefulsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149807 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.ReplicaSet: Get https://104.154.103.101/apis/apps/v1beta2/replicasets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149886 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Secret: Get https://104.154.103.101/api/v1/secrets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.149968 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Role: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/roles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.154805 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.158486 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:49.165613 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.108145 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.RoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/rolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.120265 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.PodSecurityPolicy: Get https://104.154.103.101/apis/extensions/v1beta1/podsecuritypolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.130455 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Namespace: Get https://104.154.103.101/api/v1/namespaces?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.134251 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRole: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterroles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.149426 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ServiceAccount: Get https://104.154.103.101/api/v1/serviceaccounts?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.155033 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Pod: Get https://104.154.103.101/api/v1/pods?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.158272 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.DaemonSet: Get https://104.154.103.101/apis/apps/v1beta2/daemonsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.159290 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ReplicationController: Get https://104.154.103.101/api/v1/replicationcontrollers?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.165706 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.CronJob: Get https://104.154.103.101/apis/batch/v1beta1/cronjobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.166768 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.Deployment: Get https://104.154.103.101/apis/apps/v1beta2/deployments?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.167814 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Job: Get https://104.154.103.101/apis/batch/v1/jobs?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.168880 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Endpoints: Get https://104.154.103.101/api/v1/endpoints?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.170327 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta1.Ingress: Get https://104.154.103.101/apis/extensions/v1beta1/ingresses?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.171910 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.NetworkPolicy: Get https://104.154.103.101/apis/networking.k8s.io/v1/networkpolicies?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.179251 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Event: Get https://104.154.103.101/api/v1/events?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.180306 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.StatefulSet: Get https://104.154.103.101/apis/apps/v1beta2/statefulsets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.181642 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1beta2.ReplicaSet: Get https://104.154.103.101/apis/apps/v1beta2/replicasets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.182628 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Secret: Get https://104.154.103.101/api/v1/secrets?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.183720 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Role: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/roles?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.185095 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.ClusterRoleBinding: Get https://104.154.103.101/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.186151 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Node: Get https://104.154.103.101/api/v1/nodes?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
E0404 02:04:50.187425 1 reflector.go:205] github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:129: Failed to list *v1.Service: Get https://104.154.103.101/api/v1/services?resourceVersion=0: cluster.management.cattle.io "cluster-n9gwr" not found
``` | non_priority | deletion of gke cluster results in repeated errors logs in server relating to cluster not found rancher versions build from master steps to reproduce the problem created a gke cluster added workload deleted this cluster this results in repeated errors logs in server relating to cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list clusterrolebinding get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list node get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list service get cluster management cattle io cluster not found generic controller go clusteralertcontroller cluster clusteralert controllermanager failed with delete cluster management cattle io cluster not found generic controller go clusteralertcontroller cluster clusteralert deploment event failed with delete cluster management cattle io cluster not found generic controller go clusteralertcontroller cluster clusteralert etcd failed with delete cluster management cattle io cluster not found generic controller go clusteralertcontroller cluster clusteralert node mem failed with delete cluster management cattle io cluster not found generic controller go clusteralertcontroller cluster clusteralert scheduler failed with delete cluster management cattle io cluster not found generic controller go projectalertcontroller project projectalert workload failed with delete cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list rolebinding get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list podsecuritypolicy get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list namespace get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list clusterrole get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list serviceaccount get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list pod get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list daemonset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list replicationcontroller get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list cronjob get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list deployment get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list job get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list endpoints get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list ingress get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list networkpolicy get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list event get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list statefulset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list replicaset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list secret get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list role get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list clusterrolebinding get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list node get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list service get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list rolebinding get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list podsecuritypolicy get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list namespace get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list clusterrole get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list serviceaccount get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list pod get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list daemonset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list replicationcontroller get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list cronjob get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list deployment get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list job get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list endpoints get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list ingress get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list networkpolicy get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list event get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list statefulset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list replicaset get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list secret get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list role get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list clusterrolebinding get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list node get cluster management cattle io cluster not found reflector go github com rancher rancher vendor github com rancher norman controller generic controller go failed to list service get cluster management cattle io cluster not found | 0 |
509,143 | 14,713,799,564 | IssuesEvent | 2021-01-05 10:55:37 | wso2/devstudio-tooling-ei | https://api.github.com/repos/wso2/devstudio-tooling-ei | closed | Validations are not working when creating a new API with a swagger definition | Priority/High Severity/Major Swagger Type/Bug | **Description:**
In the Integration Studio, we can create an API with a swagger definition. However, it throws NPE if the swagger definition is not in the correct format. We need to handle this tooling itself. Also, in **Create new API** wizard we can create an API without pointing a swagger definition file. It always enables the **Finish** button in the wizard to create an API.
**Suggested Labels:**
Type/Bug | 1.0 | Validations are not working when creating a new API with a swagger definition - **Description:**
In the Integration Studio, we can create an API with a swagger definition. However, it throws NPE if the swagger definition is not in the correct format. We need to handle this tooling itself. Also, in **Create new API** wizard we can create an API without pointing a swagger definition file. It always enables the **Finish** button in the wizard to create an API.
**Suggested Labels:**
Type/Bug | priority | validations are not working when creating a new api with a swagger definition description in the integration studio we can create an api with a swagger definition however it throws npe if the swagger definition is not in the correct format we need to handle this tooling itself also in create new api wizard we can create an api without pointing a swagger definition file it always enables the finish button in the wizard to create an api suggested labels type bug | 1 |
58,588 | 24,495,964,021 | IssuesEvent | 2022-10-10 08:42:12 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Unified search]- line details expand query editor and SQL reference missing dark theme | bug Team:AppServicesSv impact:low Feature:Unified search | **Kibana version:** 8.5.0 BC2
**Elasticsearch version:** 8.5.0 BC2
**Server OS version:** darwin_x86_64
**Browser version:** chrome latest
**Browser OS version:** OS X
**Original install method (e.g. download page, yum, from source, etc.):** from staging
**Describe the bug:** When user switches to SQL in dark theme line details, expand query editor and SQL reference are missing dark theme.
<img width="1728" alt="Screen Shot 2022-10-04 at 5 19 51 PM" src="https://user-images.githubusercontent.com/7074629/193933729-368549ed-5744-44dd-bf2f-4b77fa2808b7.png">
| 1.0 | [Unified search]- line details expand query editor and SQL reference missing dark theme - **Kibana version:** 8.5.0 BC2
**Elasticsearch version:** 8.5.0 BC2
**Server OS version:** darwin_x86_64
**Browser version:** chrome latest
**Browser OS version:** OS X
**Original install method (e.g. download page, yum, from source, etc.):** from staging
**Describe the bug:** When user switches to SQL in dark theme line details, expand query editor and SQL reference are missing dark theme.
<img width="1728" alt="Screen Shot 2022-10-04 at 5 19 51 PM" src="https://user-images.githubusercontent.com/7074629/193933729-368549ed-5744-44dd-bf2f-4b77fa2808b7.png">
| non_priority | line details expand query editor and sql reference missing dark theme kibana version elasticsearch version server os version darwin browser version chrome latest browser os version os x original install method e g download page yum from source etc from staging describe the bug when user switches to sql in dark theme line details expand query editor and sql reference are missing dark theme img width alt screen shot at pm src | 0 |
52,839 | 27,800,942,100 | IssuesEvent | 2023-03-17 15:47:38 | modin-project/modin | https://api.github.com/repos/modin-project/modin | opened | Evaluate the pros and cons of lazy functions submission (via `partition.add_to_apply_calls`) | Performance 🚀 | > Do we want to make this lazy? Since `split_row_partitions` is in effect the properly partitioned dataframe, we can transform to col partitions, and then `add_to_apply_calls` the sort instead, and defer metadata materialization till it's needed?
_Originally posted by @RehanSD in https://github.com/modin-project/modin/pull/5780#discussion_r1137835188_
Modin's partition class has a feature that seems to not be used at all. The feature allows forming a queue of remote calls making it possible to combine them as a single remote call to the execution engine when the materialization is actually needed, thus reducing the number of remote calls being made to the engine and potentially optimizing the workflow:
https://github.com/modin-project/modin/blob/94b2a0e86ea631451f44d4af64ab5424e42f7764/modin/core/dataframe/pandas/partitioning/partition.py#L117-L119
We once had an attempt to use this feature (#1838) which however resulted in performance regression for certain cases and thus the changes were reverted by #2471. We never tried to re-iterate the changes and figure out the root cause of the regression.
So opening this issue as a reminder that we may want to do so as IMO it has a great potential for optimizing the way we operate with the execution engine.
| True | Evaluate the pros and cons of lazy functions submission (via `partition.add_to_apply_calls`) - > Do we want to make this lazy? Since `split_row_partitions` is in effect the properly partitioned dataframe, we can transform to col partitions, and then `add_to_apply_calls` the sort instead, and defer metadata materialization till it's needed?
_Originally posted by @RehanSD in https://github.com/modin-project/modin/pull/5780#discussion_r1137835188_
Modin's partition class has a feature that seems to not be used at all. The feature allows forming a queue of remote calls making it possible to combine them as a single remote call to the execution engine when the materialization is actually needed, thus reducing the number of remote calls being made to the engine and potentially optimizing the workflow:
https://github.com/modin-project/modin/blob/94b2a0e86ea631451f44d4af64ab5424e42f7764/modin/core/dataframe/pandas/partitioning/partition.py#L117-L119
We once had an attempt to use this feature (#1838) which however resulted in performance regression for certain cases and thus the changes were reverted by #2471. We never tried to re-iterate the changes and figure out the root cause of the regression.
So opening this issue as a reminder that we may want to do so as IMO it has a great potential for optimizing the way we operate with the execution engine.
| non_priority | evaluate the pros and cons of lazy functions submission via partition add to apply calls do we want to make this lazy since split row partitions is in effect the properly partitioned dataframe we can transform to col partitions and then add to apply calls the sort instead and defer metadata materialization till it s needed originally posted by rehansd in modin s partition class has a feature that seems to not be used at all the feature allows forming a queue of remote calls making it possible to combine them as a single remote call to the execution engine when the materialization is actually needed thus reducing the number of remote calls being made to the engine and potentially optimizing the workflow we once had an attempt to use this feature which however resulted in performance regression for certain cases and thus the changes were reverted by we never tried to re iterate the changes and figure out the root cause of the regression so opening this issue as a reminder that we may want to do so as imo it has a great potential for optimizing the way we operate with the execution engine | 0 |
298,594 | 25,837,939,961 | IssuesEvent | 2022-12-12 21:17:14 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | DISABLED test_fn_grad_linalg_lu_factor_cuda_complex128 (__main__.TestBwdGradientsCUDA) | module: autograd triaged module: flaky-tests module: linear algebra skipped module: unknown | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_grad_linalg_lu_factor_cuda_complex128&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/9289278980).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_fn_grad_linalg_lu_factor_cuda_complex128`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
cc @ezyang @albanD @zou3519 @gqchen @pearu @nikitaved @soulitzer @Lezcano @Varal7 @jianyuh @mruberry @walterddr @IvanYashchuk @xwang233 | 1.0 | DISABLED test_fn_grad_linalg_lu_factor_cuda_complex128 (__main__.TestBwdGradientsCUDA) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_grad_linalg_lu_factor_cuda_complex128&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/9289278980).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_fn_grad_linalg_lu_factor_cuda_complex128`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
cc @ezyang @albanD @zou3519 @gqchen @pearu @nikitaved @soulitzer @Lezcano @Varal7 @jianyuh @mruberry @walterddr @IvanYashchuk @xwang233 | non_priority | disabled test fn grad linalg lu factor cuda main testbwdgradientscuda platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test fn grad linalg lu factor cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs cc ezyang alband gqchen pearu nikitaved soulitzer lezcano jianyuh mruberry walterddr ivanyashchuk | 0 |
101,181 | 4,108,738,053 | IssuesEvent | 2016-06-06 17:06:32 | WordPress/meta-environment | https://api.github.com/repos/WordPress/meta-environment | opened | WordCamp.dev: Make sites HTTP-only | priority: medium type: enhancement | Having them use SSL just adds confusion for people, people they run into the certificate errors. There may be some code that expects HTTPS, but we can probably work around that. | 1.0 | WordCamp.dev: Make sites HTTP-only - Having them use SSL just adds confusion for people, people they run into the certificate errors. There may be some code that expects HTTPS, but we can probably work around that. | priority | wordcamp dev make sites http only having them use ssl just adds confusion for people people they run into the certificate errors there may be some code that expects https but we can probably work around that | 1 |
79,500 | 3,535,965,498 | IssuesEvent | 2016-01-16 22:37:52 | proveit-js/proveit | https://api.github.com/repos/proveit-js/proveit | closed | Support AFC psuedo-namespace and Draft namespace | enhancement imported Priority-Medium | _From [matthew.flaschen@gatech.edu](https://code.google.com/u/108647890027017428365/) on January 04, 2014 16:10:36_
Show ProveIt in the AFC psuedo-namespace (Wikipedia talk:Articles for creation) and the Draft namespace.
_Original issue: http://code.google.com/p/proveit-js/issues/detail?id=179_ | 1.0 | Support AFC psuedo-namespace and Draft namespace - _From [matthew.flaschen@gatech.edu](https://code.google.com/u/108647890027017428365/) on January 04, 2014 16:10:36_
Show ProveIt in the AFC psuedo-namespace (Wikipedia talk:Articles for creation) and the Draft namespace.
_Original issue: http://code.google.com/p/proveit-js/issues/detail?id=179_ | priority | support afc psuedo namespace and draft namespace from on january show proveit in the afc psuedo namespace wikipedia talk articles for creation and the draft namespace original issue | 1 |
534,992 | 15,679,991,392 | IssuesEvent | 2021-03-25 01:49:51 | Landry333/Big-Owl | https://api.github.com/repos/Landry333/Big-Owl | closed | US-52: As a supervised user, I would like to respond to an attendance supervision request | medium priority user story | Details:
- The request can be either rejected or accepted
- This is the point when the user is added to the list of users that will be supervised by the user monitoring them | 1.0 | US-52: As a supervised user, I would like to respond to an attendance supervision request - Details:
- The request can be either rejected or accepted
- This is the point when the user is added to the list of users that will be supervised by the user monitoring them | priority | us as a supervised user i would like to respond to an attendance supervision request details the request can be either rejected or accepted this is the point when the user is added to the list of users that will be supervised by the user monitoring them | 1 |
403,596 | 27,425,121,603 | IssuesEvent | 2023-03-01 19:40:05 | nasa-jpl/stellar | https://api.github.com/repos/nasa-jpl/stellar | closed | Make the distinction between stellar and react-stellar clear | documentation | Make sure we properly document that Stellar is just CSS + icons and that React-Stellar builds on top of Stellar and provides general interactive components for React. Each repository should have a list of the components from Figma Stellar and indicate which components are supported by the repo. | 1.0 | Make the distinction between stellar and react-stellar clear - Make sure we properly document that Stellar is just CSS + icons and that React-Stellar builds on top of Stellar and provides general interactive components for React. Each repository should have a list of the components from Figma Stellar and indicate which components are supported by the repo. | non_priority | make the distinction between stellar and react stellar clear make sure we properly document that stellar is just css icons and that react stellar builds on top of stellar and provides general interactive components for react each repository should have a list of the components from figma stellar and indicate which components are supported by the repo | 0 |
589,848 | 17,761,839,143 | IssuesEvent | 2021-08-29 21:00:25 | ClinGen/clincoded | https://api.github.com/repos/ClinGen/clincoded | closed | Case/Seg - guidance text for selecting evidence source type | priority: high VCI colleague request requires specification dependency | Add in guidance text for selecting evidence source type in Case/Seg tab.

| 1.0 | Case/Seg - guidance text for selecting evidence source type - Add in guidance text for selecting evidence source type in Case/Seg tab.

| priority | case seg guidance text for selecting evidence source type add in guidance text for selecting evidence source type in case seg tab | 1 |
34,265 | 14,351,492,937 | IssuesEvent | 2020-11-30 01:25:00 | Azure/azure-rest-api-specs | https://api.github.com/repos/Azure/azure-rest-api-specs | closed | SecurityInsights: multiple usages of same-named discriminator | SecurityInsights Service Attention | The discriminator `Scheduled` is both used in [`ScheduledAlertRule`](https://github.com/Azure/azure-rest-api-specs/blob/d50ea7233b80f662cb9e19883fb0083e653339ca/specification/securityinsights/resource-manager/Microsoft.SecurityInsights/preview/2019-01-01-preview/SecurityInsights.json#L8705) and [`ScheduledAlertRuleTemplate`](https://github.com/Azure/azure-rest-api-specs/blob/d50ea7233b80f662cb9e19883fb0083e653339ca/specification/securityinsights/resource-manager/Microsoft.SecurityInsights/preview/2019-01-01-preview/SecurityInsights.json#L8847).
The same problem exists for `MicrosoftSecurityIncidentCreation` and `Fusion`. The issue exists in both 2019-01-01 and 2020-01-01. | 1.0 | SecurityInsights: multiple usages of same-named discriminator - The discriminator `Scheduled` is both used in [`ScheduledAlertRule`](https://github.com/Azure/azure-rest-api-specs/blob/d50ea7233b80f662cb9e19883fb0083e653339ca/specification/securityinsights/resource-manager/Microsoft.SecurityInsights/preview/2019-01-01-preview/SecurityInsights.json#L8705) and [`ScheduledAlertRuleTemplate`](https://github.com/Azure/azure-rest-api-specs/blob/d50ea7233b80f662cb9e19883fb0083e653339ca/specification/securityinsights/resource-manager/Microsoft.SecurityInsights/preview/2019-01-01-preview/SecurityInsights.json#L8847).
The same problem exists for `MicrosoftSecurityIncidentCreation` and `Fusion`. The issue exists in both 2019-01-01 and 2020-01-01. | non_priority | securityinsights multiple usages of same named discriminator the discriminator scheduled is both used in and the same problem exists for microsoftsecurityincidentcreation and fusion the issue exists in both and | 0 |
121,868 | 10,196,645,977 | IssuesEvent | 2019-08-12 21:19:32 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: splits/largerange/size=10GiB,nodes=3 failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/a850466dfc4f8eed9e7f758a61f60b120798410f
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=splits/largerange/size=10GiB,nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1426564&tab=buildLog
```
The test failed on branch=release-19.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190807-1426564/splits/largerange/size=10GiB_nodes=3/run_1
split.go:293,split.go:297,cluster.go:2078,errgroup.go:57: failed to get range count: pq: syntax error at or near "ranges"
cluster.go:2099,split.go:317,split.go:216,test_runner.go:691: Goexit() was called
``` | 2.0 | roachtest: splits/largerange/size=10GiB,nodes=3 failed - SHA: https://github.com/cockroachdb/cockroach/commits/a850466dfc4f8eed9e7f758a61f60b120798410f
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=splits/largerange/size=10GiB,nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1426564&tab=buildLog
```
The test failed on branch=release-19.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190807-1426564/splits/largerange/size=10GiB_nodes=3/run_1
split.go:293,split.go:297,cluster.go:2078,errgroup.go:57: failed to get range count: pq: syntax error at or near "ranges"
cluster.go:2099,split.go:317,split.go:216,test_runner.go:691: Goexit() was called
``` | non_priority | roachtest splits largerange size nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests splits largerange size nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts splits largerange size nodes run split go split go cluster go errgroup go failed to get range count pq syntax error at or near ranges cluster go split go split go test runner go goexit was called | 0 |
429,681 | 30,085,235,840 | IssuesEvent | 2023-06-29 08:08:43 | bokeh/dataviz-fundamentals | https://api.github.com/repos/bokeh/dataviz-fundamentals | closed | Introduction files | tag: documentation tag: notebook | The `Introduction.ipynb` and `Introduction.html` files till contain the old file paths. I have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts. | 1.0 | Introduction files - The `Introduction.ipynb` and `Introduction.html` files till contain the old file paths. I have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts. | non_priority | introduction files the introduction ipynb and introduction html files till contain the old file paths i have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts | 0 |
53,943 | 23,115,105,131 | IssuesEvent | 2022-07-27 15:58:23 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Revert automated Data Tracker Fulcrum updates | Workgroup: AMD Service: Apps Need: 2-Should Have | <!-- Email -->
<!-- billy.bolander@austintexas.gov -->
> What application are you using?
Other / Not Sure
> Describe the problem.
At this time Fulcrum is updating Tracker when a PM has been completed. This was by my request in the past. Due to a change in our internal procedure I request this to be reversed. Please disallow Fulcrum to update Tracker for PM's and allow the closing of the PM work order to be the update for the PM completed date.
> How soon do you need this?
Flexible — An extended timeline is OK
> Is there anything else we should know?
I will monitor these entries until this request can be filled
> Requested By
Billy B.
Request ID: DTS20-101295
| 1.0 | Revert automated Data Tracker Fulcrum updates - <!-- Email -->
<!-- billy.bolander@austintexas.gov -->
> What application are you using?
Other / Not Sure
> Describe the problem.
At this time Fulcrum is updating Tracker when a PM has been completed. This was by my request in the past. Due to a change in our internal procedure I request this to be reversed. Please disallow Fulcrum to update Tracker for PM's and allow the closing of the PM work order to be the update for the PM completed date.
> How soon do you need this?
Flexible — An extended timeline is OK
> Is there anything else we should know?
I will monitor these entries until this request can be filled
> Requested By
Billy B.
Request ID: DTS20-101295
| non_priority | revert automated data tracker fulcrum updates what application are you using other not sure describe the problem at this time fulcrum is updating tracker when a pm has been completed this was by my request in the past due to a change in our internal procedure i request this to be reversed please disallow fulcrum to update tracker for pm s and allow the closing of the pm work order to be the update for the pm completed date how soon do you need this flexible — an extended timeline is ok is there anything else we should know i will monitor these entries until this request can be filled requested by billy b request id | 0 |
408,003 | 11,940,532,178 | IssuesEvent | 2020-04-02 16:52:21 | eclipse-ee4j/jaxb-api | https://api.github.com/repos/eclipse-ee4j/jaxb-api | closed | XMLGregorianCalendar needs a getXmlSchemaType() method | Component: datatypes Priority: Major Type: Improvement | Submitted by David Bau, entered initially by Joe Fialli
There should be a method to say "which one of the XML Schema Date
types does this correspond to"? based on the set fields, e.g., have an
enumeration for gYear, gMonth, gMonthYear, as well as no-type, and then have
a getXmlSchemaType() method. What's the use case? We have some customers
with schemas where they have unions of various XML Schema Date types, and it
is very useful to know which member of the union they have.
#### Environment
Operating System: All
Platform: All
#### Affected Versions
[2.0 draft] | 1.0 | XMLGregorianCalendar needs a getXmlSchemaType() method - Submitted by David Bau, entered initially by Joe Fialli
There should be a method to say "which one of the XML Schema Date
types does this correspond to"? based on the set fields, e.g., have an
enumeration for gYear, gMonth, gMonthYear, as well as no-type, and then have
a getXmlSchemaType() method. What's the use case? We have some customers
with schemas where they have unions of various XML Schema Date types, and it
is very useful to know which member of the union they have.
#### Environment
Operating System: All
Platform: All
#### Affected Versions
[2.0 draft] | priority | xmlgregoriancalendar needs a getxmlschematype method submitted by david bau entered initially by joe fialli there should be a method to say which one of the xml schema date types does this correspond to based on the set fields e g have an enumeration for gyear gmonth gmonthyear as well as no type and then have a getxmlschematype method what s the use case we have some customers with schemas where they have unions of various xml schema date types and it is very useful to know which member of the union they have environment operating system all platform all affected versions | 1 |
60,698 | 8,454,311,895 | IssuesEvent | 2018-10-21 01:17:28 | Naoghuman/app-notification | https://api.github.com/repos/Naoghuman/app-notification | closed | [doc] Create a basic layout concept for the application gui. | documentation | [doc] Create a basic layout concept for the application gui.
* The main gui is primaly a TabPane where the user can configure the single notification topics. | 1.0 | [doc] Create a basic layout concept for the application gui. - [doc] Create a basic layout concept for the application gui.
* The main gui is primaly a TabPane where the user can configure the single notification topics. | non_priority | create a basic layout concept for the application gui create a basic layout concept for the application gui the main gui is primaly a tabpane where the user can configure the single notification topics | 0 |
80,193 | 23,138,188,430 | IssuesEvent | 2022-07-28 15:54:11 | Kotlin/multik | https://api.github.com/repos/Kotlin/multik | opened | _concat_fortran_string | build native | deal with the dependence to calculate eigenvalues and vectors. get rid of `quadmath` dependency and build for all platforms. | 1.0 | _concat_fortran_string - deal with the dependence to calculate eigenvalues and vectors. get rid of `quadmath` dependency and build for all platforms. | non_priority | concat fortran string deal with the dependence to calculate eigenvalues and vectors get rid of quadmath dependency and build for all platforms | 0 |
300,841 | 22,699,540,417 | IssuesEvent | 2022-07-05 09:24:17 | AlbertoCuadra/combustion_toolbox | https://api.github.com/repos/AlbertoCuadra/combustion_toolbox | closed | Validation: include validations in the docs - part I | documentation validation | - [x] Test TP: Equilibrium composition at defined T and p
- [x] Test HP: Adiabatic T and composition at constant p
- [x] Test SHOCK_R: Planar reflected shock wave
- [x] Test DET: Chapman-Jouget Detonation (CJ upper state)
- [x] Test IONIZATION
- [x] Test SHOCK_POLAR | 1.0 | Validation: include validations in the docs - part I - - [x] Test TP: Equilibrium composition at defined T and p
- [x] Test HP: Adiabatic T and composition at constant p
- [x] Test SHOCK_R: Planar reflected shock wave
- [x] Test DET: Chapman-Jouget Detonation (CJ upper state)
- [x] Test IONIZATION
- [x] Test SHOCK_POLAR | non_priority | validation include validations in the docs part i test tp equilibrium composition at defined t and p test hp adiabatic t and composition at constant p test shock r planar reflected shock wave test det chapman jouget detonation cj upper state test ionization test shock polar | 0 |
595,698 | 18,073,301,882 | IssuesEvent | 2021-09-21 06:55:11 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.beatport.com - site is not usable | os-ios browser-firefox-ios priority-normal device-tablet | <!-- @browser: Firefox -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/87255 -->
<!-- @extra_labels: browser-firefox-ios, device-tablet -->
**URL**: https://www.beatport.com/hold-bin/tracks
**Browser / Version**: Firefox
**Operating System**: I Pad 14.6
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Der Play Butten funktioniert nicht
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/9/04a30a97-61d9-453c-b93f-da28d35f95bb.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.beatport.com - site is not usable - <!-- @browser: Firefox -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/87255 -->
<!-- @extra_labels: browser-firefox-ios, device-tablet -->
**URL**: https://www.beatport.com/hold-bin/tracks
**Browser / Version**: Firefox
**Operating System**: I Pad 14.6
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Der Play Butten funktioniert nicht
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/9/04a30a97-61d9-453c-b93f-da28d35f95bb.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox operating system i pad tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce der play butten funktioniert nicht view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
15,483 | 19,693,135,301 | IssuesEvent | 2022-01-12 09:23:03 | bisq-network/bisq | https://api.github.com/repos/bisq-network/bisq | closed | Improve user experience once mediation has been accepted by both parties | in:gui a:feature in:trade-process | <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team over at https://keybase.io/team/Bisq
-->
### Description
When mediation for a given trade is accepted by both parties the trade moves from 'open trades' to 'history' and the payout transaction ID is broadcast.
This is how it should be but the user experience could be achieved by keeping the trade in 'open trades' for longer and give the user the option to withdraw funds to Bisq wallet or external wallet. This would add an additional step that is the same as what happens currently when trades in 'open trades' are successfully completed ie, the user is notified and then they make the choice to keep funds in Bisq wallet or withdraw them. This would improve the user experience consistency.
#### Version
v 1.6.5
### Steps to reproduce
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is no notification. Next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out. Trader 1 will need to check funds > transactions to see the status of the trade.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is not informed mediation has been accepted, there is no notification. The trade will have moved from open trades to history and the payout tx id will be broadcast.
### Expected behaviour
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is a notification. The mediated trade will still be in open trades and the Trader 1 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is informed mediation has been accepted, by a notification. The trade will still be in open trades and Trader 2 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
| 1.0 | Improve user experience once mediation has been accepted by both parties - <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team over at https://keybase.io/team/Bisq
-->
### Description
When mediation for a given trade is accepted by both parties the trade moves from 'open trades' to 'history' and the payout transaction ID is broadcast.
This is how it should be but the user experience could be achieved by keeping the trade in 'open trades' for longer and give the user the option to withdraw funds to Bisq wallet or external wallet. This would add an additional step that is the same as what happens currently when trades in 'open trades' are successfully completed ie, the user is notified and then they make the choice to keep funds in Bisq wallet or withdraw them. This would improve the user experience consistency.
#### Version
v 1.6.5
### Steps to reproduce
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is no notification. Next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out. Trader 1 will need to check funds > transactions to see the status of the trade.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is not informed mediation has been accepted, there is no notification. The trade will have moved from open trades to history and the payout tx id will be broadcast.
### Expected behaviour
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is a notification. The mediated trade will still be in open trades and the Trader 1 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is informed mediation has been accepted, by a notification. The trade will still be in open trades and Trader 2 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
| non_priority | improve user experience once mediation has been accepted by both parties support requests this is for reporting bugs in the bisq app if you have a support request please join support on bisq s keybase team over at description when mediation for a given trade is accepted by both parties the trade moves from open trades to history and the payout transaction id is broadcast this is how it should be but the user experience could be achieved by keeping the trade in open trades for longer and give the user the option to withdraw funds to bisq wallet or external wallet this would add an additional step that is the same as what happens currently when trades in open trades are successfully completed ie the user is notified and then they make the choice to keep funds in bisq wallet or withdraw them this would improve the user experience consistency version v steps to reproduce from the perspective of trader initial trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is no notification next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out trader will need to check funds transactions to see the status of the trade from the perspective of trader second trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is no notification the trade will have moved from open trades to history and the payout tx id will be broadcast expected behaviour from the perspective of trader initial trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is a notification the mediated trade will still be in open trades and the trader will have the option to keep the funds in their bisq wallet or withdraw to an external wallet the trade will then move to history from the perspective of trader second trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is informed mediation has been accepted by a notification the trade will still be in open trades and trader will have the option to keep the funds in their bisq wallet or withdraw to an external wallet the trade will then move to history | 0 |
683,499 | 23,384,600,029 | IssuesEvent | 2022-08-11 12:45:19 | TheYellowArchitect/doubledamnation | https://api.github.com/repos/TheYellowArchitect/doubledamnation | opened | Remake Enemy Behaviour | enemy ai low priority | It is no secret that the code for the enemies is **not** well-designed. After all, it was made back in 2018, I didn't even know a design pattern back then.
I didn't read on AI, it was quite literally handmade without any guidance, I still remember the "breakthroughs" and how making the AI felt like exploring, happy times.
There are some good things, like the split of enemy onto pathfinder and behaviour, and the attacks being split in classes, hence monsters are "modular" (e.g. hollow with some clicks, could get the jump attack of satyr, or the rush attack of centaur and minotaur)
However! There are 3 horrible things, which linger to this very day, as I didn't know anything about AI, and I was coding whatever came into my head:
1. **Attack Detection Hitbox and Attack Damage Hitbox are MERGED**
Yes, this is as horrible as it sounds. The same hitbox which says "if in X range, use your first attack" is also the hitbox which checks for applying damage within X range.
It isn't really noticeable in-game, because with the above, enemies attack when they get in range of their attack. But it is noticeable for Harpy's spin attack.
Harpy's spinattack is clunky/janky, because of this flawed design.
2. **Sidewall Bug**
There was a weird bug I cannot recall, when the AI was adjacent to a wall. Related to satyr, probably (this was back in late 2018, dont blame me)
To "fix" this, I did a sidewall check. Which did "work" but I didn't see at the time that it influenced other behaviour.
While I am certain the fix is simple, just refreshing myself on that codebase and what the original bug was, will take a week, and this is a chore.
It is a serious bug, however, because enemies at wall sides have bugged AI.
3. Flying Type
Instead of splitting EnemyBehaviour into Grounded and Flying, I slapped Flying code onto default EnemyBehaviour. And if (flying) then bloated code overrides default.
Not a bug, not wrong, but it makes the code ugly and bloated. | 1.0 | Remake Enemy Behaviour - It is no secret that the code for the enemies is **not** well-designed. After all, it was made back in 2018, I didn't even know a design pattern back then.
I didn't read on AI, it was quite literally handmade without any guidance, I still remember the "breakthroughs" and how making the AI felt like exploring, happy times.
There are some good things, like the split of enemy onto pathfinder and behaviour, and the attacks being split in classes, hence monsters are "modular" (e.g. hollow with some clicks, could get the jump attack of satyr, or the rush attack of centaur and minotaur)
However! There are 3 horrible things, which linger to this very day, as I didn't know anything about AI, and I was coding whatever came into my head:
1. **Attack Detection Hitbox and Attack Damage Hitbox are MERGED**
Yes, this is as horrible as it sounds. The same hitbox which says "if in X range, use your first attack" is also the hitbox which checks for applying damage within X range.
It isn't really noticeable in-game, because with the above, enemies attack when they get in range of their attack. But it is noticeable for Harpy's spin attack.
Harpy's spinattack is clunky/janky, because of this flawed design.
2. **Sidewall Bug**
There was a weird bug I cannot recall, when the AI was adjacent to a wall. Related to satyr, probably (this was back in late 2018, dont blame me)
To "fix" this, I did a sidewall check. Which did "work" but I didn't see at the time that it influenced other behaviour.
While I am certain the fix is simple, just refreshing myself on that codebase and what the original bug was, will take a week, and this is a chore.
It is a serious bug, however, because enemies at wall sides have bugged AI.
3. Flying Type
Instead of splitting EnemyBehaviour into Grounded and Flying, I slapped Flying code onto default EnemyBehaviour. And if (flying) then bloated code overrides default.
Not a bug, not wrong, but it makes the code ugly and bloated. | priority | remake enemy behaviour it is no secret that the code for the enemies is not well designed after all it was made back in i didn t even know a design pattern back then i didn t read on ai it was quite literally handmade without any guidance i still remember the breakthroughs and how making the ai felt like exploring happy times there are some good things like the split of enemy onto pathfinder and behaviour and the attacks being split in classes hence monsters are modular e g hollow with some clicks could get the jump attack of satyr or the rush attack of centaur and minotaur however there are horrible things which linger to this very day as i didn t know anything about ai and i was coding whatever came into my head attack detection hitbox and attack damage hitbox are merged yes this is as horrible as it sounds the same hitbox which says if in x range use your first attack is also the hitbox which checks for applying damage within x range it isn t really noticeable in game because with the above enemies attack when they get in range of their attack but it is noticeable for harpy s spin attack harpy s spinattack is clunky janky because of this flawed design sidewall bug there was a weird bug i cannot recall when the ai was adjacent to a wall related to satyr probably this was back in late dont blame me to fix this i did a sidewall check which did work but i didn t see at the time that it influenced other behaviour while i am certain the fix is simple just refreshing myself on that codebase and what the original bug was will take a week and this is a chore it is a serious bug however because enemies at wall sides have bugged ai flying type instead of splitting enemybehaviour into grounded and flying i slapped flying code onto default enemybehaviour and if flying then bloated code overrides default not a bug not wrong but it makes the code ugly and bloated | 1 |
327,130 | 24,119,379,419 | IssuesEvent | 2022-09-20 17:16:28 | The-Mycelium-Network/bookclub | https://api.github.com/repos/The-Mycelium-Network/bookclub | opened | chore: add some notes for chapter 1 of JavaScript the definitive guide | documentation chore | I have some notes I want to add for the first chapter of JavaScript, the definitive guide | 1.0 | chore: add some notes for chapter 1 of JavaScript the definitive guide - I have some notes I want to add for the first chapter of JavaScript, the definitive guide | non_priority | chore add some notes for chapter of javascript the definitive guide i have some notes i want to add for the first chapter of javascript the definitive guide | 0 |
153,817 | 5,903,993,436 | IssuesEvent | 2017-05-19 08:36:15 | metasfresh/metasfresh | https://api.github.com/repos/metasfresh/metasfresh | closed | HU Transform - split out some TUs from LU does not work correctly with custom LUs | priority:high release:candidate status:integrated type:bug | ### Is this a bug or feature request?
Bug
### What is the current behavior?
#### Which are the steps to reproduce?
Consider following setup and test case (you can have it with different numbers):
* have a product which has the CU-TU capacity: 7 Kg per TU
* receive an LU with 10 TUs each of the TU shall have 10 Kg (and NOT 7Kg as the standard says)
* use HU Transform process to take out 2 TUs
You will get:
* 3 TUs instead of 2 TUs
** TU1 : 7Kg
** TU2: 7Kg
** TU3: 6Kg
That's because the BL considered the standard TU capacity when splitting out.
But when the BL calculated how much Qty CU in total shall be splitted out, it used 2x10Kg=20Kg
### What is the expected or desired behavior?
We shall get 2 TUs, each of them having 10Kg.
| 1.0 | HU Transform - split out some TUs from LU does not work correctly with custom LUs - ### Is this a bug or feature request?
Bug
### What is the current behavior?
#### Which are the steps to reproduce?
Consider following setup and test case (you can have it with different numbers):
* have a product which has the CU-TU capacity: 7 Kg per TU
* receive an LU with 10 TUs each of the TU shall have 10 Kg (and NOT 7Kg as the standard says)
* use HU Transform process to take out 2 TUs
You will get:
* 3 TUs instead of 2 TUs
** TU1 : 7Kg
** TU2: 7Kg
** TU3: 6Kg
That's because the BL considered the standard TU capacity when splitting out.
But when the BL calculated how much Qty CU in total shall be splitted out, it used 2x10Kg=20Kg
### What is the expected or desired behavior?
We shall get 2 TUs, each of them having 10Kg.
| priority | hu transform split out some tus from lu does not work correctly with custom lus is this a bug or feature request bug what is the current behavior which are the steps to reproduce consider following setup and test case you can have it with different numbers have a product which has the cu tu capacity kg per tu receive an lu with tus each of the tu shall have kg and not as the standard says use hu transform process to take out tus you will get tus instead of tus that s because the bl considered the standard tu capacity when splitting out but when the bl calculated how much qty cu in total shall be splitted out it used what is the expected or desired behavior we shall get tus each of them having | 1 |
329,120 | 24,209,219,680 | IssuesEvent | 2022-09-25 17:05:19 | pha4ge/hAMRonization | https://api.github.com/repos/pha4ge/hAMRonization | closed | Update README | documentation enhancement p3 | - [ ] Improve installation instructions
- [ ] Update parsers included
- [ ] Add wiki with small tutorial | 1.0 | Update README - - [ ] Improve installation instructions
- [ ] Update parsers included
- [ ] Add wiki with small tutorial | non_priority | update readme improve installation instructions update parsers included add wiki with small tutorial | 0 |
487,892 | 14,061,025,708 | IssuesEvent | 2020-11-03 07:20:35 | bryntum/support | https://api.github.com/repos/bryntum/support | closed | Export to MSP - Add support for version MSP 2019 | bug high-priority resolved | Reference: https://github.com/bryntum/support/issues/1250
- [x] Make work for version MSP 2019
- [x] Export Assignments with percent work complete
- [x] Export CalendarUID for resources
Version 2013:

Version 2019:

XML File Generated:
[Launch SaaS Product.xml.zip](https://github.com/bryntum/support/files/5295305/Launch.SaaS.Product.xml.zip)
| 1.0 | Export to MSP - Add support for version MSP 2019 - Reference: https://github.com/bryntum/support/issues/1250
- [x] Make work for version MSP 2019
- [x] Export Assignments with percent work complete
- [x] Export CalendarUID for resources
Version 2013:

Version 2019:

XML File Generated:
[Launch SaaS Product.xml.zip](https://github.com/bryntum/support/files/5295305/Launch.SaaS.Product.xml.zip)
| priority | export to msp add support for version msp reference make work for version msp export assignments with percent work complete export calendaruid for resources version version xml file generated | 1 |
70,000 | 13,392,515,249 | IssuesEvent | 2020-09-03 01:35:06 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | XCode 12 beta "Your Xcode version may be too old for your iOS version." when running iOS 14 | P1 customer: crowd e: OS Version specific platform-ios severe: regression t: xcode tool | ## Steps to Reproduce
1. create hello word app.
2. run `flutter run` or `flutter run --release` on iPhone X with iOS 14
**Expected results:** run app perfectly.
**Actual results:** getting error `Your Xcode version may be too old for your iOS version.`
it is ok when run with xCode but also show white screen for almost 5 minutes in startup | 1.0 | XCode 12 beta "Your Xcode version may be too old for your iOS version." when running iOS 14 - ## Steps to Reproduce
1. create hello word app.
2. run `flutter run` or `flutter run --release` on iPhone X with iOS 14
**Expected results:** run app perfectly.
**Actual results:** getting error `Your Xcode version may be too old for your iOS version.`
it is ok when run with xCode but also show white screen for almost 5 minutes in startup | non_priority | xcode beta your xcode version may be too old for your ios version when running ios steps to reproduce create hello word app run flutter run or flutter run release on iphone x with ios expected results run app perfectly actual results getting error your xcode version may be too old for your ios version it is ok when run with xcode but also show white screen for almost minutes in startup | 0 |
216,815 | 24,309,079,629 | IssuesEvent | 2022-09-29 20:17:19 | opensearch-project/security | https://api.github.com/repos/opensearch-project/security | closed | CVE-2022-34917 (High) detected in kafka-clients-3.0.1.jar - autoclosed | security vulnerability | ## CVE-2022-34917 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kafka-clients-3.0.1.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://kafka.apache.org">https://kafka.apache.org</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.kafka/kafka-clients/3.0.1/8f931e45e96e952728d540829e5bde9d79fab172/kafka-clients-3.0.1.jar,/radle/caches/modules-2/files-2.1/org.apache.kafka/kafka-clients/3.0.1/8f931e45e96e952728d540829e5bde9d79fab172/kafka-clients-3.0.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **kafka-clients-3.0.1.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security vulnerability has been identified in Apache Kafka. It affects all releases since 2.8.0. The vulnerability allows malicious unauthenticated clients to allocate large amounts of memory on brokers. This can lead to brokers hitting OutOfMemoryException and causing denial of service. Example scenarios: - Kafka cluster without authentication: Any clients able to establish a network connection to a broker can trigger the issue. - Kafka cluster with SASL authentication: Any clients able to establish a network connection to a broker, without the need for valid SASL credentials, can trigger the issue. - Kafka cluster with TLS authentication: Only clients able to successfully authenticate via TLS can trigger the issue. We advise the users to upgrade the Kafka installations to one of the 3.2.3, 3.1.2, 3.0.2, 2.8.2 versions.
<p>Publish Date: 2022-09-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34917>CVE-2022-34917</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-34917">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-34917</a></p>
<p>Release Date: 2022-09-20</p>
<p>Fix Resolution: 3.0.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2022-34917 (High) detected in kafka-clients-3.0.1.jar - autoclosed - ## CVE-2022-34917 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kafka-clients-3.0.1.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://kafka.apache.org">https://kafka.apache.org</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.kafka/kafka-clients/3.0.1/8f931e45e96e952728d540829e5bde9d79fab172/kafka-clients-3.0.1.jar,/radle/caches/modules-2/files-2.1/org.apache.kafka/kafka-clients/3.0.1/8f931e45e96e952728d540829e5bde9d79fab172/kafka-clients-3.0.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **kafka-clients-3.0.1.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security vulnerability has been identified in Apache Kafka. It affects all releases since 2.8.0. The vulnerability allows malicious unauthenticated clients to allocate large amounts of memory on brokers. This can lead to brokers hitting OutOfMemoryException and causing denial of service. Example scenarios: - Kafka cluster without authentication: Any clients able to establish a network connection to a broker can trigger the issue. - Kafka cluster with SASL authentication: Any clients able to establish a network connection to a broker, without the need for valid SASL credentials, can trigger the issue. - Kafka cluster with TLS authentication: Only clients able to successfully authenticate via TLS can trigger the issue. We advise the users to upgrade the Kafka installations to one of the 3.2.3, 3.1.2, 3.0.2, 2.8.2 versions.
<p>Publish Date: 2022-09-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34917>CVE-2022-34917</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-34917">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-34917</a></p>
<p>Release Date: 2022-09-20</p>
<p>Fix Resolution: 3.0.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_priority | cve high detected in kafka clients jar autoclosed cve high severity vulnerability vulnerable library kafka clients jar library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org apache kafka kafka clients kafka clients jar radle caches modules files org apache kafka kafka clients kafka clients jar dependency hierarchy x kafka clients jar vulnerable library found in base branch main vulnerability details a security vulnerability has been identified in apache kafka it affects all releases since the vulnerability allows malicious unauthenticated clients to allocate large amounts of memory on brokers this can lead to brokers hitting outofmemoryexception and causing denial of service example scenarios kafka cluster without authentication any clients able to establish a network connection to a broker can trigger the issue kafka cluster with sasl authentication any clients able to establish a network connection to a broker without the need for valid sasl credentials can trigger the issue kafka cluster with tls authentication only clients able to successfully authenticate via tls can trigger the issue we advise the users to upgrade the kafka installations to one of the versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr | 0 |
189,943 | 22,047,158,449 | IssuesEvent | 2022-05-30 04:01:00 | madhans23/linux-4.1.15 | https://api.github.com/repos/madhans23/linux-4.1.15 | closed | CVE-2017-12153 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed | security vulnerability | ## CVE-2017-12153 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/wireless/nl80211.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security flaw was discovered in the nl80211_set_rekey_data() function in net/wireless/nl80211.c in the Linux kernel through 4.13.3. This function does not check whether the required attributes are present in a Netlink request. This request can be issued by a user with the CAP_NET_ADMIN capability and may result in a NULL pointer dereference and system crash.
<p>Publish Date: 2017-09-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-12153>CVE-2017-12153</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-12153">https://www.linuxkernelcves.com/cves/CVE-2017-12153</a></p>
<p>Release Date: 2017-09-21</p>
<p>Fix Resolution: v4.14-rc2,v3.16.49,v3.18.73,v3.2.94,v4.1.46,v4.13.5,v4.4.90,v4.9.53</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-12153 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2017-12153 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/wireless/nl80211.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security flaw was discovered in the nl80211_set_rekey_data() function in net/wireless/nl80211.c in the Linux kernel through 4.13.3. This function does not check whether the required attributes are present in a Netlink request. This request can be issued by a user with the CAP_NET_ADMIN capability and may result in a NULL pointer dereference and system crash.
<p>Publish Date: 2017-09-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-12153>CVE-2017-12153</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-12153">https://www.linuxkernelcves.com/cves/CVE-2017-12153</a></p>
<p>Release Date: 2017-09-21</p>
<p>Fix Resolution: v4.14-rc2,v3.16.49,v3.18.73,v3.2.94,v4.1.46,v4.13.5,v4.4.90,v4.9.53</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net wireless c vulnerability details a security flaw was discovered in the set rekey data function in net wireless c in the linux kernel through this function does not check whether the required attributes are present in a netlink request this request can be issued by a user with the cap net admin capability and may result in a null pointer dereference and system crash publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
20,886 | 10,556,689,872 | IssuesEvent | 2019-10-04 03:01:36 | istio/istio | https://api.github.com/repos/istio/istio | closed | Secure coding & testing practices that affect all new work -code-mauve style | area/security kind/need more info | (NOTE: This is used to report product bugs:
To report a security vulnerability, please visit <https://istio.io/about/security-vulnerabilities/>
To ask questions about how to use Istio, please visit <https://discuss.istio.io>
)
**Bug description**
This task is a spin off of #13618 , the goal here might be to verify the state of golangci linter, see if we're using the latest linters, and if all codebase (including older) passes through linting without errors.
Also we might want to investigate possible secure coding practice not covered by existing linting.
As for the testing, this is TBD (was not explicitly described in original issue).
**Affected product area (please put an X in all that apply)**
[ ] Configuration Infrastructure
[ ] Docs
[ ] Installation
[ ] Networking
[ ] Performance and Scalability
[ ] Policies and Telemetry
[X] Security
[X] Test and Release
[ ] User Experience
[ ] Developer Infrastructure
**Expected behavior**
**Steps to reproduce the bug**
**Version (include the output of `istioctl version --remote` and `kubectl version`)**
**How was Istio installed?**
**Environment where bug was observed (cloud vendor, OS, etc)**
Additionally, please consider attaching a [cluster state archive](http://istio.io/help/bugs/#generating-a-cluster-state-archive) by attaching
the dump file to this issue.
| True | Secure coding & testing practices that affect all new work -code-mauve style - (NOTE: This is used to report product bugs:
To report a security vulnerability, please visit <https://istio.io/about/security-vulnerabilities/>
To ask questions about how to use Istio, please visit <https://discuss.istio.io>
)
**Bug description**
This task is a spin off of #13618 , the goal here might be to verify the state of golangci linter, see if we're using the latest linters, and if all codebase (including older) passes through linting without errors.
Also we might want to investigate possible secure coding practice not covered by existing linting.
As for the testing, this is TBD (was not explicitly described in original issue).
**Affected product area (please put an X in all that apply)**
[ ] Configuration Infrastructure
[ ] Docs
[ ] Installation
[ ] Networking
[ ] Performance and Scalability
[ ] Policies and Telemetry
[X] Security
[X] Test and Release
[ ] User Experience
[ ] Developer Infrastructure
**Expected behavior**
**Steps to reproduce the bug**
**Version (include the output of `istioctl version --remote` and `kubectl version`)**
**How was Istio installed?**
**Environment where bug was observed (cloud vendor, OS, etc)**
Additionally, please consider attaching a [cluster state archive](http://istio.io/help/bugs/#generating-a-cluster-state-archive) by attaching
the dump file to this issue.
| non_priority | secure coding testing practices that affect all new work code mauve style note this is used to report product bugs to report a security vulnerability please visit to ask questions about how to use istio please visit bug description this task is a spin off of the goal here might be to verify the state of golangci linter see if we re using the latest linters and if all codebase including older passes through linting without errors also we might want to investigate possible secure coding practice not covered by existing linting as for the testing this is tbd was not explicitly described in original issue affected product area please put an x in all that apply configuration infrastructure docs installation networking performance and scalability policies and telemetry security test and release user experience developer infrastructure expected behavior steps to reproduce the bug version include the output of istioctl version remote and kubectl version how was istio installed environment where bug was observed cloud vendor os etc additionally please consider attaching a by attaching the dump file to this issue | 0 |
72,130 | 8,705,492,248 | IssuesEvent | 2018-12-05 22:36:08 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | closed | Blow out destination options default, source well, and dest well | feature medium protocol designer | Related to issue #1676
As a user I'd like better control over where blowout occurs. Currently my options are source plate A1, destination, trash.
As a user I'd like a surfaced default selection, so I can't select blow out and then nothing in the destination dropdown
## Acceptance Criteria
- [ ] Blow out destination options are: (1) Source well, (2) Destination well, (3) Trash
- [ ] Blow out default displays in dropdown. There's no null state. | 1.0 | Blow out destination options default, source well, and dest well - Related to issue #1676
As a user I'd like better control over where blowout occurs. Currently my options are source plate A1, destination, trash.
As a user I'd like a surfaced default selection, so I can't select blow out and then nothing in the destination dropdown
## Acceptance Criteria
- [ ] Blow out destination options are: (1) Source well, (2) Destination well, (3) Trash
- [ ] Blow out default displays in dropdown. There's no null state. | non_priority | blow out destination options default source well and dest well related to issue as a user i d like better control over where blowout occurs currently my options are source plate destination trash as a user i d like a surfaced default selection so i can t select blow out and then nothing in the destination dropdown acceptance criteria blow out destination options are source well destination well trash blow out default displays in dropdown there s no null state | 0 |
93,019 | 26,837,688,595 | IssuesEvent | 2023-02-02 20:59:21 | uselagoon/lagoon | https://api.github.com/repos/uselagoon/lagoon | closed | Invalid YAML in pvc annotations | bug 2-build-deploy | # Describe the bug
The YAML generated by Helm is invalid for `pvc` in the `nginx-php-persistent` template
```yaml
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: nginx-php
labels:
helm.sh/chart: nginx-php-persistent-0.1.0
app.kubernetes.io/name: nginx-php-persistent
app.kubernetes.io/instance: nginx-php
app.kubernetes.io/managed-by: Helm
lagoon.sh/service: nginx-php
lagoon.sh/service-type: nginx-php-persistent
lagoon.sh/project: REDACTED
lagoon.sh/environment: production
lagoon.sh/environmentType: production
lagoon.sh/buildType: branch
annotations:
lagoon.sh/version: "21.2.2"
lagoon.sh/branch: "production"
annotations:
k8up.syn.tools/backup: "true"
spec:
accessModes:
- ReadWriteMany
storageClassName: bulk
resources:
requests:
storage: "20Gi"
```
Notice the double `annotations` key.
## To Reproduce
Do a deploy, notice the the YAML is not valid.
## Expected behavior
Valid YAML being generated by Helm
| 1.0 | Invalid YAML in pvc annotations - # Describe the bug
The YAML generated by Helm is invalid for `pvc` in the `nginx-php-persistent` template
```yaml
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: nginx-php
labels:
helm.sh/chart: nginx-php-persistent-0.1.0
app.kubernetes.io/name: nginx-php-persistent
app.kubernetes.io/instance: nginx-php
app.kubernetes.io/managed-by: Helm
lagoon.sh/service: nginx-php
lagoon.sh/service-type: nginx-php-persistent
lagoon.sh/project: REDACTED
lagoon.sh/environment: production
lagoon.sh/environmentType: production
lagoon.sh/buildType: branch
annotations:
lagoon.sh/version: "21.2.2"
lagoon.sh/branch: "production"
annotations:
k8up.syn.tools/backup: "true"
spec:
accessModes:
- ReadWriteMany
storageClassName: bulk
resources:
requests:
storage: "20Gi"
```
Notice the double `annotations` key.
## To Reproduce
Do a deploy, notice the the YAML is not valid.
## Expected behavior
Valid YAML being generated by Helm
| non_priority | invalid yaml in pvc annotations describe the bug the yaml generated by helm is invalid for pvc in the nginx php persistent template yaml apiversion kind persistentvolumeclaim metadata name nginx php labels helm sh chart nginx php persistent app kubernetes io name nginx php persistent app kubernetes io instance nginx php app kubernetes io managed by helm lagoon sh service nginx php lagoon sh service type nginx php persistent lagoon sh project redacted lagoon sh environment production lagoon sh environmenttype production lagoon sh buildtype branch annotations lagoon sh version lagoon sh branch production annotations syn tools backup true spec accessmodes readwritemany storageclassname bulk resources requests storage notice the double annotations key to reproduce do a deploy notice the the yaml is not valid expected behavior valid yaml being generated by helm | 0 |
197,066 | 6,952,042,980 | IssuesEvent | 2017-12-06 16:14:56 | vmware/vic-product | https://api.github.com/repos/vmware/vic-product | opened | Clarify NFS docs | component/vic-engine kind/user-doc priority/high pub/vsphere | Per Slack chat with @zjs, @jooskim, @matthewavery:
- The API is just a passthrough to the library, but the UI seems to have a dropdown for selecting a datastore (which is a good thing), but no way to turn that into a free-form text box to supply an NFS server. I think it's an easy thing to add post-1.3
- nfs volumes should show up
- in my dev environment i have one nfs volume attached to a cluster, for which i get an entry under the dropdown
- we haven’t made significant changes to the way we fetch datastores so it should be showing up
- NFS datastores show up, but only in the same way that other datastores do. It looks like there's a lot more flexibility when configuring NFS on the command-line:
- so i thought that once you’ve selected a cluster, one of the hosts under the cluster should have the nfs volume available
- NFS datastores show up, but there's no way to specify an NFS server that isn't a datastore or configure NFS-specific settings (which might be a gap in the API as well).
- no the nfs volume store does not need to be backed by an nfs datastore. An NFS datastore is actually a datastore that is backed by a sharepoint.
- Sounds like we'll have to file issues to fix this in the API and CLI, and have a note directing the user to the command-line for 1.3
- For this reason, the nfs volumestore target is not intended to be a datastore(atleast for now). But rather a sharepoint. If you do use an nfs datastore as the target then we will make vmdks on it, which are not shareable.
So, for 1.3, we need to document that if your NFS sharepoint is not an NFS datastore, you need to do `vic-machine configure` after deployment, to add it. @zjs @matthewavery is this accurate? | 1.0 | Clarify NFS docs - Per Slack chat with @zjs, @jooskim, @matthewavery:
- The API is just a passthrough to the library, but the UI seems to have a dropdown for selecting a datastore (which is a good thing), but no way to turn that into a free-form text box to supply an NFS server. I think it's an easy thing to add post-1.3
- nfs volumes should show up
- in my dev environment i have one nfs volume attached to a cluster, for which i get an entry under the dropdown
- we haven’t made significant changes to the way we fetch datastores so it should be showing up
- NFS datastores show up, but only in the same way that other datastores do. It looks like there's a lot more flexibility when configuring NFS on the command-line:
- so i thought that once you’ve selected a cluster, one of the hosts under the cluster should have the nfs volume available
- NFS datastores show up, but there's no way to specify an NFS server that isn't a datastore or configure NFS-specific settings (which might be a gap in the API as well).
- no the nfs volume store does not need to be backed by an nfs datastore. An NFS datastore is actually a datastore that is backed by a sharepoint.
- Sounds like we'll have to file issues to fix this in the API and CLI, and have a note directing the user to the command-line for 1.3
- For this reason, the nfs volumestore target is not intended to be a datastore(atleast for now). But rather a sharepoint. If you do use an nfs datastore as the target then we will make vmdks on it, which are not shareable.
So, for 1.3, we need to document that if your NFS sharepoint is not an NFS datastore, you need to do `vic-machine configure` after deployment, to add it. @zjs @matthewavery is this accurate? | priority | clarify nfs docs per slack chat with zjs jooskim matthewavery the api is just a passthrough to the library but the ui seems to have a dropdown for selecting a datastore which is a good thing but no way to turn that into a free form text box to supply an nfs server i think it s an easy thing to add post nfs volumes should show up in my dev environment i have one nfs volume attached to a cluster for which i get an entry under the dropdown we haven’t made significant changes to the way we fetch datastores so it should be showing up nfs datastores show up but only in the same way that other datastores do it looks like there s a lot more flexibility when configuring nfs on the command line so i thought that once you’ve selected a cluster one of the hosts under the cluster should have the nfs volume available nfs datastores show up but there s no way to specify an nfs server that isn t a datastore or configure nfs specific settings which might be a gap in the api as well no the nfs volume store does not need to be backed by an nfs datastore an nfs datastore is actually a datastore that is backed by a sharepoint sounds like we ll have to file issues to fix this in the api and cli and have a note directing the user to the command line for for this reason the nfs volumestore target is not intended to be a datastore atleast for now but rather a sharepoint if you do use an nfs datastore as the target then we will make vmdks on it which are not shareable so for we need to document that if your nfs sharepoint is not an nfs datastore you need to do vic machine configure after deployment to add it zjs matthewavery is this accurate | 1 |
274,051 | 23,806,529,684 | IssuesEvent | 2022-09-04 05:37:25 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: schemachange/mixed-versions failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.schemachange/mixed-versions [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6341855?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6341855?buildTab=artifacts#/schemachange/mixed-versions) on master @ [b0f13c9bbef3e6628471d887672be7c7658f6511](https://github.com/cockroachdb/cockroach/commits/b0f13c9bbef3e6628471d887672be7c7658f6511):
```
| 300.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 301.0s 1 0.0 0.5 0.0 0.0 0.0 0.0 opOk
| 301.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 301.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 301.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 302.0s 1 0.0 0.5 0.0 0.0 0.0 0.0 opOk
| 302.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 302.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 302.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 303.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 303.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 303.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 303.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 304.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 304.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 304.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 304.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 305.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 305.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 305.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 305.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 306.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 306.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 306.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 306.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 307.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 307.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 307.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 307.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 308.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 308.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 308.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 308.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 309.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 309.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 309.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 309.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 310.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 310.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 310.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 310.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
Wraps: (4) SSH_PROBLEM
Wraps: (5) Node 3. Command with error:
| ``````
| ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4}
| ``````
Wraps: (6) exit status 255
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.SSH (5) *hintdetail.withDetail (6) *exec.ExitError
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: schemachange/mixed-versions failed - roachtest.schemachange/mixed-versions [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6341855?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6341855?buildTab=artifacts#/schemachange/mixed-versions) on master @ [b0f13c9bbef3e6628471d887672be7c7658f6511](https://github.com/cockroachdb/cockroach/commits/b0f13c9bbef3e6628471d887672be7c7658f6511):
```
| 300.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 301.0s 1 0.0 0.5 0.0 0.0 0.0 0.0 opOk
| 301.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 301.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 301.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 302.0s 1 0.0 0.5 0.0 0.0 0.0 0.0 opOk
| 302.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 302.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 302.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 303.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 303.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 303.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 303.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 304.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 304.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 304.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 304.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 305.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 305.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 305.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 305.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 306.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 306.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 306.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 306.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 307.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 307.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 307.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 307.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 308.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 308.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 308.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 308.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 309.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 309.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 309.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 309.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
| 310.0s 1 0.0 0.4 0.0 0.0 0.0 0.0 opOk
| 310.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnCmtErr
| 310.0s 1 0.0 0.1 0.0 0.0 0.0 0.0 txnOk
| 310.0s 1 0.0 0.0 0.0 0.0 0.0 0.0 txnRbk
Wraps: (4) SSH_PROBLEM
Wraps: (5) Node 3. Command with error:
| ``````
| ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4}
| ``````
Wraps: (6) exit status 255
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.SSH (5) *hintdetail.withDetail (6) *exec.ExitError
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_priority | roachtest schemachange mixed versions failed roachtest schemachange mixed versions with on master txnrbk opok txncmterr txnok elapsed errors ops sec inst ops sec cum ms ms ms pmax ms txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok elapsed errors ops sec inst ops sec cum ms ms ms pmax ms txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk opok txncmterr txnok txnrbk wraps ssh problem wraps node command with error workload run schemachange verbose tolerate errors true max ops concurrency pgurl wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors ssh hintdetail withdetail exec exiterror parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb sql schema | 0 |
321,404 | 23,853,006,552 | IssuesEvent | 2022-09-06 19:52:09 | Requisitos-de-Software/2022.1-TikTok | https://api.github.com/repos/Requisitos-de-Software/2022.1-TikTok | closed | Atualizar versionamento no documento de rich picture | documentation | ## Descrição
Adicionar datas no versionamento do documento de rich picture
## Tarefas
- [x] Adicionar datas
## Critérios de aceitação
- [x] Documento atualizado no repositório | 1.0 | Atualizar versionamento no documento de rich picture - ## Descrição
Adicionar datas no versionamento do documento de rich picture
## Tarefas
- [x] Adicionar datas
## Critérios de aceitação
- [x] Documento atualizado no repositório | non_priority | atualizar versionamento no documento de rich picture descrição adicionar datas no versionamento do documento de rich picture tarefas adicionar datas critérios de aceitação documento atualizado no repositório | 0 |
23,907 | 16,680,735,138 | IssuesEvent | 2021-06-07 23:11:37 | votingworks/vxsuite | https://api.github.com/repos/votingworks/vxsuite | closed | Share `jest.config.js` settings | [zube]: In Progress infrastructure | This can likely be accomplished with a `jest.shared.config.js` at the root that the others just `require` and override as needed. | 1.0 | Share `jest.config.js` settings - This can likely be accomplished with a `jest.shared.config.js` at the root that the others just `require` and override as needed. | non_priority | share jest config js settings this can likely be accomplished with a jest shared config js at the root that the others just require and override as needed | 0 |
92,006 | 8,336,264,068 | IssuesEvent | 2018-09-28 07:08:38 | owncloud/core | https://api.github.com/repos/owncloud/core | closed | Add automated phpcs checks in drone for acceptance test code | QA-team dev:acceptance-tests | The acceptance test code has been conforming to a proposed "standard" that was implemented some time ago but not formally enforced. It requires a bunch of PHP doc block standards, for example. See ``phpcs.xml`` in the root folder of the core repo.
The check is done with:
```
phpcs --standard=phpcs.xml tests/acceptance
```
Implement this much checking as a drone job, so that at least the acceptance test code stays to this standard.
Note: It is out-of-scope for this issue to decide what standard might apply to the whole code base and how to implement such a thing.
| 1.0 | Add automated phpcs checks in drone for acceptance test code - The acceptance test code has been conforming to a proposed "standard" that was implemented some time ago but not formally enforced. It requires a bunch of PHP doc block standards, for example. See ``phpcs.xml`` in the root folder of the core repo.
The check is done with:
```
phpcs --standard=phpcs.xml tests/acceptance
```
Implement this much checking as a drone job, so that at least the acceptance test code stays to this standard.
Note: It is out-of-scope for this issue to decide what standard might apply to the whole code base and how to implement such a thing.
| non_priority | add automated phpcs checks in drone for acceptance test code the acceptance test code has been conforming to a proposed standard that was implemented some time ago but not formally enforced it requires a bunch of php doc block standards for example see phpcs xml in the root folder of the core repo the check is done with phpcs standard phpcs xml tests acceptance implement this much checking as a drone job so that at least the acceptance test code stays to this standard note it is out of scope for this issue to decide what standard might apply to the whole code base and how to implement such a thing | 0 |
251,222 | 8,001,531,157 | IssuesEvent | 2018-07-23 03:41:56 | alipay/sofa-mosn | https://api.github.com/repos/alipay/sofa-mosn | closed | route wildcard domain match rule: longest wildcard suffix match | area/router kind/bug priority/P1 status/done | two questions:
1.
golang map's iteration order is not specified and is not guaranteed to be the same from one iteration to the next.
to longest wildcard suffix match, we need to sort key in wildcardVirtualHostSuffixes.
2.
Only unique values for domains are permitted, even if it is wildcard domain.
| 1.0 | route wildcard domain match rule: longest wildcard suffix match - two questions:
1.
golang map's iteration order is not specified and is not guaranteed to be the same from one iteration to the next.
to longest wildcard suffix match, we need to sort key in wildcardVirtualHostSuffixes.
2.
Only unique values for domains are permitted, even if it is wildcard domain.
| priority | route wildcard domain match rule longest wildcard suffix match two questions golang map s iteration order is not specified and is not guaranteed to be the same from one iteration to the next to longest wildcard suffix match we need to sort key in wildcardvirtualhostsuffixes only unique values for domains are permitted even if it is wildcard domain | 1 |
190,698 | 6,821,649,267 | IssuesEvent | 2017-11-07 17:23:12 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | Implement IP lookup for VCH create endpoint | area/ui priority/high team/lifecycle | **Details:**
We need to implement the findings from this issue https://github.com/vmware/vic/issues/6551 regarding the H5C VCH wizard being able find the correct IP of vic-machine-service.
**Acceptance Criteria:**
Demonstrate ability to deploy OVA with vic-machine-service and have the H5C vch wizard be able to find the correct API endpoint and get a successful response.
| 1.0 | Implement IP lookup for VCH create endpoint - **Details:**
We need to implement the findings from this issue https://github.com/vmware/vic/issues/6551 regarding the H5C VCH wizard being able find the correct IP of vic-machine-service.
**Acceptance Criteria:**
Demonstrate ability to deploy OVA with vic-machine-service and have the H5C vch wizard be able to find the correct API endpoint and get a successful response.
| priority | implement ip lookup for vch create endpoint details we need to implement the findings from this issue regarding the vch wizard being able find the correct ip of vic machine service acceptance criteria demonstrate ability to deploy ova with vic machine service and have the vch wizard be able to find the correct api endpoint and get a successful response | 1 |
35,950 | 9,691,079,929 | IssuesEvent | 2019-05-24 10:12:00 | Lundalogik/lip | https://api.github.com/repos/Lundalogik/lip | opened | Format lip.json so that it looks properly formatted | enhancement package builder | Today, everything ends up in one loooong line.
Write to several lines and indent it properly. | 1.0 | Format lip.json so that it looks properly formatted - Today, everything ends up in one loooong line.
Write to several lines and indent it properly. | non_priority | format lip json so that it looks properly formatted today everything ends up in one loooong line write to several lines and indent it properly | 0 |
91,284 | 3,851,440,429 | IssuesEvent | 2016-04-06 02:02:45 | cs2103jan2016-w10-3j/main | https://api.github.com/repos/cs2103jan2016-w10-3j/main | closed | A user can be alerted when events are close to expiration | priority.medium type.story | So the user wouldn't forget to do them. | 1.0 | A user can be alerted when events are close to expiration - So the user wouldn't forget to do them. | priority | a user can be alerted when events are close to expiration so the user wouldn t forget to do them | 1 |
100,798 | 30,777,243,885 | IssuesEvent | 2023-07-31 07:31:11 | SigNoz/signoz | https://api.github.com/repos/SigNoz/signoz | closed | FE(Dashboard): Migrate current dashboard api into v3 query range | enhancement frontend query-builder | ## Is your feature request related to a problem?
*Please describe.*
Right now in the dashboard we are using v2 we need to migrate it to v3 | 1.0 | FE(Dashboard): Migrate current dashboard api into v3 query range - ## Is your feature request related to a problem?
*Please describe.*
Right now in the dashboard we are using v2 we need to migrate it to v3 | non_priority | fe dashboard migrate current dashboard api into query range is your feature request related to a problem please describe right now in the dashboard we are using we need to migrate it to | 0 |
776,595 | 27,264,466,529 | IssuesEvent | 2023-02-22 16:59:41 | ascheid/itsg33-pbmm-issue-gen | https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen | opened | AC-20: Use Of External Information Systems | Priority: P2 Class: Technical Suggested Assignment: IT Security Function ITSG-33 Control: AC-20 | # Control Definition
(A) The organization establishes terms and conditions, consistent with any trust relationships established with other organizations owning, operating, and/or maintaining external information systems, allowing authorized individuals to access the information system from external information systems.
(B) The organization establishes terms and conditions, consistent with any trust relationships established with other organizations owning, operating, and/or maintaining external information systems, allowing authorized individuals to process, store, or transmit organization-controlled information using external information systems.
# Class
Technical
# Supplemental Guidance
External information systems are information systems or components of information systems that are outside of the authorization boundary established by organizations and for which organizations typically have no direct supervision and authority over the application of required security controls or the assessment of control effectiveness. External information systems include, for example: (i) personally owned information systems/devices (e.g., notebook computers, smart phones, tablets, personal digital assistants); (ii) privately owned computing and communications devices resident in commercial or public facilities (e.g., hotels, train stations, convention centres, shopping malls, or airports); (iii) information systems owned or controlled by non-GC organizations; and (iv) GC information systems that are not owned by, operated by, or under the direct supervision and authority of organizations. This control also addresses the use of external information systems for the processing, storage, or transmission of organizational information, including, for example, accessing cloud services (e.g., infrastructure as a service, platform as a service, or software as a service) from organizational information systemsFor some external information systems (i.e., information systems operated by other GC agencies, including organizations subordinate to those GC organizations), the trust relationships that have been established between those organizations and the originating organization may be such, that no explicit terms and conditions are required. Information systems within these organizations would not be considered external. These situations occur when, for example, there are pre-existing sharing/trust agreements (either implicit or explicit) established between GC organizations subordinate to those GC organizations, or when such trust agreements are specified by applicable GC legislation and TBS policies, directives and standards. Authorized individuals include, for example, organizational personnel, contractors, or other individuals with authorized access to organizational information systems and over which organizations have the authority to impose rules of behaviour with regard to system access. Restrictions that organizations impose on authorized individuals need not be uniform, as those restrictions may vary depending upon the trust relationships between organizations. Therefore, organizations may choose to impose different security restrictions on contractors than on federal, provincial, or municipal government.This control does not apply to the use of external information systems to access public interfaces to organizational information systems. Organizations establish terms and conditions for the use of external information systems in accordance with organizational security policies and procedures. Terms and conditions address as a minimum: types of applications that can be accessed on organizational information systems from external information systems; and the highest security category of information that can be processed, stored, or transmitted on external information systems. If terms and conditions with the owners of external information systems cannot be established, organizations may impose restrictions on organizational personnel using those external systems. Related controls: AC-3, AC-17, AC-19, CA-3, PL-4, SA-9.
# General Guide
This security control/enhancement is considered to be best practice. Consequently, inclusion in a departmental profile is strongly encouraged in most cases.
# Suggested Assignment
IT Security Function
| 1.0 | AC-20: Use Of External Information Systems - # Control Definition
(A) The organization establishes terms and conditions, consistent with any trust relationships established with other organizations owning, operating, and/or maintaining external information systems, allowing authorized individuals to access the information system from external information systems.
(B) The organization establishes terms and conditions, consistent with any trust relationships established with other organizations owning, operating, and/or maintaining external information systems, allowing authorized individuals to process, store, or transmit organization-controlled information using external information systems.
# Class
Technical
# Supplemental Guidance
External information systems are information systems or components of information systems that are outside of the authorization boundary established by organizations and for which organizations typically have no direct supervision and authority over the application of required security controls or the assessment of control effectiveness. External information systems include, for example: (i) personally owned information systems/devices (e.g., notebook computers, smart phones, tablets, personal digital assistants); (ii) privately owned computing and communications devices resident in commercial or public facilities (e.g., hotels, train stations, convention centres, shopping malls, or airports); (iii) information systems owned or controlled by non-GC organizations; and (iv) GC information systems that are not owned by, operated by, or under the direct supervision and authority of organizations. This control also addresses the use of external information systems for the processing, storage, or transmission of organizational information, including, for example, accessing cloud services (e.g., infrastructure as a service, platform as a service, or software as a service) from organizational information systemsFor some external information systems (i.e., information systems operated by other GC agencies, including organizations subordinate to those GC organizations), the trust relationships that have been established between those organizations and the originating organization may be such, that no explicit terms and conditions are required. Information systems within these organizations would not be considered external. These situations occur when, for example, there are pre-existing sharing/trust agreements (either implicit or explicit) established between GC organizations subordinate to those GC organizations, or when such trust agreements are specified by applicable GC legislation and TBS policies, directives and standards. Authorized individuals include, for example, organizational personnel, contractors, or other individuals with authorized access to organizational information systems and over which organizations have the authority to impose rules of behaviour with regard to system access. Restrictions that organizations impose on authorized individuals need not be uniform, as those restrictions may vary depending upon the trust relationships between organizations. Therefore, organizations may choose to impose different security restrictions on contractors than on federal, provincial, or municipal government.This control does not apply to the use of external information systems to access public interfaces to organizational information systems. Organizations establish terms and conditions for the use of external information systems in accordance with organizational security policies and procedures. Terms and conditions address as a minimum: types of applications that can be accessed on organizational information systems from external information systems; and the highest security category of information that can be processed, stored, or transmitted on external information systems. If terms and conditions with the owners of external information systems cannot be established, organizations may impose restrictions on organizational personnel using those external systems. Related controls: AC-3, AC-17, AC-19, CA-3, PL-4, SA-9.
# General Guide
This security control/enhancement is considered to be best practice. Consequently, inclusion in a departmental profile is strongly encouraged in most cases.
# Suggested Assignment
IT Security Function
| priority | ac use of external information systems control definition a the organization establishes terms and conditions consistent with any trust relationships established with other organizations owning operating and or maintaining external information systems allowing authorized individuals to access the information system from external information systems b the organization establishes terms and conditions consistent with any trust relationships established with other organizations owning operating and or maintaining external information systems allowing authorized individuals to process store or transmit organization controlled information using external information systems class technical supplemental guidance external information systems are information systems or components of information systems that are outside of the authorization boundary established by organizations and for which organizations typically have no direct supervision and authority over the application of required security controls or the assessment of control effectiveness external information systems include for example i personally owned information systems devices e g notebook computers smart phones tablets personal digital assistants ii privately owned computing and communications devices resident in commercial or public facilities e g hotels train stations convention centres shopping malls or airports iii information systems owned or controlled by non gc organizations and iv gc information systems that are not owned by operated by or under the direct supervision and authority of organizations this control also addresses the use of external information systems for the processing storage or transmission of organizational information including for example accessing cloud services e g infrastructure as a service platform as a service or software as a service from organizational information systemsfor some external information systems i e information systems operated by other gc agencies including organizations subordinate to those gc organizations the trust relationships that have been established between those organizations and the originating organization may be such that no explicit terms and conditions are required information systems within these organizations would not be considered external these situations occur when for example there are pre existing sharing trust agreements either implicit or explicit established between gc organizations subordinate to those gc organizations or when such trust agreements are specified by applicable gc legislation and tbs policies directives and standards authorized individuals include for example organizational personnel contractors or other individuals with authorized access to organizational information systems and over which organizations have the authority to impose rules of behaviour with regard to system access restrictions that organizations impose on authorized individuals need not be uniform as those restrictions may vary depending upon the trust relationships between organizations therefore organizations may choose to impose different security restrictions on contractors than on federal provincial or municipal government this control does not apply to the use of external information systems to access public interfaces to organizational information systems organizations establish terms and conditions for the use of external information systems in accordance with organizational security policies and procedures terms and conditions address as a minimum types of applications that can be accessed on organizational information systems from external information systems and the highest security category of information that can be processed stored or transmitted on external information systems if terms and conditions with the owners of external information systems cannot be established organizations may impose restrictions on organizational personnel using those external systems related controls ac ac ac ca pl sa general guide this security control enhancement is considered to be best practice consequently inclusion in a departmental profile is strongly encouraged in most cases suggested assignment it security function | 1 |
241,404 | 20,120,647,589 | IssuesEvent | 2022-02-08 01:40:57 | free-tomorrow/free-tomorrow | https://api.github.com/repos/free-tomorrow/free-tomorrow | closed | FE: Cypress Iteration 1: | feature FE testing | Write tests to confirm that:
- [ ] The header elements display correctly, and any links redirect to the correct Routes
- [ ] The homepage elements display correctly and any buttons redirect to the correct Routes | 1.0 | FE: Cypress Iteration 1: - Write tests to confirm that:
- [ ] The header elements display correctly, and any links redirect to the correct Routes
- [ ] The homepage elements display correctly and any buttons redirect to the correct Routes | non_priority | fe cypress iteration write tests to confirm that the header elements display correctly and any links redirect to the correct routes the homepage elements display correctly and any buttons redirect to the correct routes | 0 |
780,854 | 27,411,010,206 | IssuesEvent | 2023-03-01 10:31:25 | horizon-efrei/HorizonBot | https://api.github.com/repos/horizon-efrei/HorizonBot | closed | Feature pour délégués: TODO-list de devoirs | type: feature difficulty: complex status: awaiting approval scope: class groups priority: lowest | Ensemble de commandes pour gérer de façon semi-automatique une liste de devoirs à faire pour les délégués dans les classes
Commandes à prévoir: !todo create <liste des matières>, !todo add <matière> <intitulé> <date due>, !todo remove <matière/intitulé>, !todo edit <intitulé>, !todo archive
Exemple de layout (devoirs à faire):
--- DEVOIRS A FAIRE (S8) ----
**Natural Language Processing (NLP): (projet terminé ❌ DE Passé ❌)**
- Amazon TP Analysis (01/02)
- TP1: Analysis of Presidential Speeches (05/02)
- Envoyer un HTML/PDF du Notebook (avant le 06/02)
Exemple de layout archive:
--- DEVOIRS FAITS (S7) ----
**Big Data Frameworks: (projet terminé :ballot_box_with_check: DE Passé :ballot_box_with_check: )**
- TP4 Lab1 avec Markdown (15 Oct. 23:59 sur Moodle)
- TP4 Lab2 (Yarn/JAVA) (12 Nov. 23:59 sur Moodle)
- TP5 (Hive) (15 Nov. 23:59 sur Moodle)
- TP6 (HBase) (23 Nov, pas de rendu) | 1.0 | Feature pour délégués: TODO-list de devoirs - Ensemble de commandes pour gérer de façon semi-automatique une liste de devoirs à faire pour les délégués dans les classes
Commandes à prévoir: !todo create <liste des matières>, !todo add <matière> <intitulé> <date due>, !todo remove <matière/intitulé>, !todo edit <intitulé>, !todo archive
Exemple de layout (devoirs à faire):
--- DEVOIRS A FAIRE (S8) ----
**Natural Language Processing (NLP): (projet terminé ❌ DE Passé ❌)**
- Amazon TP Analysis (01/02)
- TP1: Analysis of Presidential Speeches (05/02)
- Envoyer un HTML/PDF du Notebook (avant le 06/02)
Exemple de layout archive:
--- DEVOIRS FAITS (S7) ----
**Big Data Frameworks: (projet terminé :ballot_box_with_check: DE Passé :ballot_box_with_check: )**
- TP4 Lab1 avec Markdown (15 Oct. 23:59 sur Moodle)
- TP4 Lab2 (Yarn/JAVA) (12 Nov. 23:59 sur Moodle)
- TP5 (Hive) (15 Nov. 23:59 sur Moodle)
- TP6 (HBase) (23 Nov, pas de rendu) | priority | feature pour délégués todo list de devoirs ensemble de commandes pour gérer de façon semi automatique une liste de devoirs à faire pour les délégués dans les classes commandes à prévoir todo create todo add todo remove todo edit todo archive exemple de layout devoirs à faire devoirs a faire natural language processing nlp projet terminé ❌ de passé ❌ amazon tp analysis analysis of presidential speeches envoyer un html pdf du notebook avant le exemple de layout archive devoirs faits big data frameworks projet terminé ballot box with check de passé ballot box with check avec markdown oct sur moodle yarn java nov sur moodle hive nov sur moodle hbase nov pas de rendu | 1 |
249,654 | 26,968,376,070 | IssuesEvent | 2023-02-09 01:14:25 | turkdevops/icu | https://api.github.com/repos/turkdevops/icu | opened | cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl: 1 vulnerabilities (highest severity is: 4.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /tools/commit-checker/Pipfile</p>
<p>Path to vulnerable library: /tools/commit-checker/Pipfile</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cryptography version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl | Direct | cryptography - 39.0.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary>
### Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /tools/commit-checker/Pipfile</p>
<p>Path to vulnerable library: /tools/commit-checker/Pipfile</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>gh-pages</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8.
<p>Publish Date: 2023-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p>
<p>Release Date: 2023-02-07</p>
<p>Fix Resolution: cryptography - 39.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl: 1 vulnerabilities (highest severity is: 4.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /tools/commit-checker/Pipfile</p>
<p>Path to vulnerable library: /tools/commit-checker/Pipfile</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cryptography version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl | Direct | cryptography - 39.0.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary>
### Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /tools/commit-checker/Pipfile</p>
<p>Path to vulnerable library: /tools/commit-checker/Pipfile</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>gh-pages</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8.
<p>Publish Date: 2023-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p>
<p>Release Date: 2023-02-07</p>
<p>Fix Resolution: cryptography - 39.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_priority | cryptography whl vulnerabilities highest severity is vulnerable library cryptography whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file tools commit checker pipfile path to vulnerable library tools commit checker pipfile vulnerabilities cve severity cvss dependency type fixed in cryptography version remediation available medium cryptography whl direct cryptography details cve vulnerable library cryptography whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file tools commit checker pipfile path to vulnerable library tools commit checker pipfile dependency hierarchy x cryptography whl vulnerable library found in base branch gh pages vulnerability details cryptography is a package designed to expose cryptographic primitives and recipes to python developers in affected versions cipher update into would accept python objects which implement the buffer protocol but provide only immutable buffers this would allow immutable objects such as bytes to be mutated thus violating fundamental rules of python and resulting in corrupted output this now correctly raises an exception this issue has been present since update into was originally introduced in cryptography publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptography step up your open source security game with mend | 0 |
44,324 | 5,796,349,355 | IssuesEvent | 2017-05-02 19:14:35 | scrisenbery/dotfiles | https://api.github.com/repos/scrisenbery/dotfiles | closed | Brew directory configuration | area/apps area/install branch/master issue/new OS/mac type/bug type/design type/question | From #1
gmac has brew all done in $HOME. The gmac in particular needs this but it probably doesn't make sense on the general scale to use that config for other computers. Will open new issue for gmac-specific branch as well.
| 1.0 | Brew directory configuration - From #1
gmac has brew all done in $HOME. The gmac in particular needs this but it probably doesn't make sense on the general scale to use that config for other computers. Will open new issue for gmac-specific branch as well.
| non_priority | brew directory configuration from gmac has brew all done in home the gmac in particular needs this but it probably doesn t make sense on the general scale to use that config for other computers will open new issue for gmac specific branch as well | 0 |
250,614 | 7,979,146,271 | IssuesEvent | 2018-07-17 20:41:58 | conveyal/analysis-ui | https://api.github.com/repos/conveyal/analysis-ui | closed | Support email should not be hard-coded | low priority small task | We may want to specify the support email in the same place we specify API keys, instead of https://github.com/conveyal/analysis-ui/blob/c9a53e9f74386de40450d5d9e4d3b0f852b14b0a/lib/components/application.js#L245
This would make sure that our support email address only shows up for our supported deployments. | 1.0 | Support email should not be hard-coded - We may want to specify the support email in the same place we specify API keys, instead of https://github.com/conveyal/analysis-ui/blob/c9a53e9f74386de40450d5d9e4d3b0f852b14b0a/lib/components/application.js#L245
This would make sure that our support email address only shows up for our supported deployments. | priority | support email should not be hard coded we may want to specify the support email in the same place we specify api keys instead of this would make sure that our support email address only shows up for our supported deployments | 1 |
101,705 | 4,128,779,295 | IssuesEvent | 2016-06-10 08:16:04 | pixelhumain/communecter | https://api.github.com/repos/pixelhumain/communecter | closed | Remove resized usage => prefere profilThumbImageUrl on every object | enhancement priority 2 | @oceatoon @Kgneo @clement59
The use of the resizer is deprecated.
We can now use the profilThumbImageUrl property on every object.
This property is set on new profil image or with new objects.
With old objects, the thumb is not generated yet. Change the image to have it.
Keep it in mind when you see a resizer call : change it ! | 1.0 | Remove resized usage => prefere profilThumbImageUrl on every object - @oceatoon @Kgneo @clement59
The use of the resizer is deprecated.
We can now use the profilThumbImageUrl property on every object.
This property is set on new profil image or with new objects.
With old objects, the thumb is not generated yet. Change the image to have it.
Keep it in mind when you see a resizer call : change it ! | priority | remove resized usage prefere profilthumbimageurl on every object oceatoon kgneo the use of the resizer is deprecated we can now use the profilthumbimageurl property on every object this property is set on new profil image or with new objects with old objects the thumb is not generated yet change the image to have it keep it in mind when you see a resizer call change it | 1 |
470,732 | 13,543,433,083 | IssuesEvent | 2020-09-16 18:57:16 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :212426] Unrecoverable parse warning in drivers/wifi/eswifi/eswifi_socket_offload.c | Coverity bug priority: low |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/66bd06a7d1f9e4682faafbc551046af695fa1060/drivers/wifi/eswifi/eswifi_socket_offload.c#L502
Category: Parse warnings
Function: ``
Component: Drivers
CID: [212426](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=212426)
Details:
```
496 (const struct fd_op_vtable *)
497 &eswifi_socket_fd_op_vtable);
498
499 return fd;
500 }
501
>>> CID 212426: Parse warnings (PARSE_ERROR)
>>> function "eswifi_socket_close" has already been defined
502 static int eswifi_socket_close(void *obj)
503 {
504 return eswifi_socket_close(sd);
505 }
506
507 static int eswifi_socket_ioctl(void *obj, unsigned int request, va_list args)
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID :212426] Unrecoverable parse warning in drivers/wifi/eswifi/eswifi_socket_offload.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/66bd06a7d1f9e4682faafbc551046af695fa1060/drivers/wifi/eswifi/eswifi_socket_offload.c#L502
Category: Parse warnings
Function: ``
Component: Drivers
CID: [212426](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=212426)
Details:
```
496 (const struct fd_op_vtable *)
497 &eswifi_socket_fd_op_vtable);
498
499 return fd;
500 }
501
>>> CID 212426: Parse warnings (PARSE_ERROR)
>>> function "eswifi_socket_close" has already been defined
502 static int eswifi_socket_close(void *obj)
503 {
504 return eswifi_socket_close(sd);
505 }
506
507 static int eswifi_socket_ioctl(void *obj, unsigned int request, va_list args)
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | unrecoverable parse warning in drivers wifi eswifi eswifi socket offload c static code scan issues found in file category parse warnings function component drivers cid details const struct fd op vtable eswifi socket fd op vtable return fd cid parse warnings parse error function eswifi socket close has already been defined static int eswifi socket close void obj return eswifi socket close sd static int eswifi socket ioctl void obj unsigned int request va list args please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
67,190 | 12,886,307,491 | IssuesEvent | 2020-07-13 09:17:00 | khochaynhalam-dev/khochaynhalam-dev.github.io | https://api.github.com/repos/khochaynhalam-dev/khochaynhalam-dev.github.io | closed | Build - [khochaynhalam] - Build Pages | code enhancement | Dear @khochaynhalam-dev/khochaynhalam-team
Task Build Pages contains the following subtasks:
- [ ] #2
- [ ] #3
- [ ] #4
- [ ] #5
- [ ] #7
Please help me do it.
Thanks,
TrungNhan
| 1.0 | Build - [khochaynhalam] - Build Pages - Dear @khochaynhalam-dev/khochaynhalam-team
Task Build Pages contains the following subtasks:
- [ ] #2
- [ ] #3
- [ ] #4
- [ ] #5
- [ ] #7
Please help me do it.
Thanks,
TrungNhan
| non_priority | build build pages dear khochaynhalam dev khochaynhalam team task build pages contains the following subtasks please help me do it thanks trungnhan | 0 |
218,262 | 7,330,874,043 | IssuesEvent | 2018-03-05 11:27:04 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Exclude httpclient_4.3.1.wso2v2.jar from the product and add the updated version : httpclient_4.3.6.wso2v1.jar | 2.2.0 Priority/Highest Resolution/Fixed Type/Task | **Description:**
Release an **openid4java** orbit bundle by fixing the import version range and exclude httpclient_4.3.1.wso2v2.jar from the product.
And then remove the exclusion of httpclient_4.3.6.wso2v1.jar to make it available in the product to make available the latest fixes to the httpclient.
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | 1.0 | Exclude httpclient_4.3.1.wso2v2.jar from the product and add the updated version : httpclient_4.3.6.wso2v1.jar - **Description:**
Release an **openid4java** orbit bundle by fixing the import version range and exclude httpclient_4.3.1.wso2v2.jar from the product.
And then remove the exclusion of httpclient_4.3.6.wso2v1.jar to make it available in the product to make available the latest fixes to the httpclient.
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | priority | exclude httpclient jar from the product and add the updated version httpclient jar description release an orbit bundle by fixing the import version range and exclude httpclient jar from the product and then remove the exclusion of httpclient jar to make it available in the product to make available the latest fixes to the httpclient suggested labels suggested assignees affected product version os db other environment details and versions steps to reproduce related issues | 1 |
806,219 | 29,806,536,787 | IssuesEvent | 2023-06-16 12:08:16 | penrose/penrose | https://api.github.com/repos/penrose/penrose | opened | feat: Make it possible to "bake" diagrams | kind:enhancement system:optimization system:style kind:usability priority:feature request layout priority:low | ### Issue
Penrose uses random sampling to initialize variables (e.g., those marked `?` in a Style program, or shape properties that are not explicitly specified). If a user is happy with a particular instance of a diagram, they would like to be able to reproduce this diagram later down the road. For this reason, we introduced at some point the notion of a _variation_, i.e., a string that is used to determine the random seed prior to diagram generation. However, variations do not completely solve the problem of reproducibility, since changes to the system itself (e.g,. the compiler or optimizer) can also change the appearance of the final diagram. Hence, it would be nice to have a more "iron clad" solution for finalizing diagrams.
One answer is to simply say: the final diagram can be stored as an SVG. However, saving as SVG does not preserve editability of non-optimized features of the diagrams (e.g., colors, line widths, text labels, etc.).
### Proposal
Instead, this issue raises the following proposal:
> Add functionality to "bake" diagrams, by writing a Style file where all optimized attributes are explicitly overriden with the optimized constants.
For instance, suppose we have the programs
```
-- Domain
type Circle
```
```
-- Substance
Circle c1
```
```
-- Style
canvas {
width = 100
height = 100
}
forall Circle C {
shape C.icon = Circle {
r: 25
center: (?,?)
}
}
```
and the optimizer yields a center `(-20.21, 15.45)` for the circle. Then to bake this diagram, we would append the following lines to the Style program:
```
forall Circle `c1` {
override `c1`.icon.center.x = -20.21
override `c1`.icon.center.y = 15.45
}
```
This feature could be exposed by an additional IDE button like `Bake` that downloads the augmented Style, and a command line option `--bake-style` that likewise dumps the augmented Style.
### Implementation
There's probably some grungy stuff that would have to happen to implement this feature, but the core thing is just getting a list of (i) optimized variables, (ii) their paths in the Style program, and (iii) their final, optimized values, which shouldn't be too hard to get our hands on.
### Field-specific baking
An elaboration on this design would be to allow the Style file to specify properties that should be baked, using the usual selector paradigm. E.g., one could write
```
forall Circle `c1` {
bake `c1`.icon.center.x
bake `c1`.icon.center.y
}
```
to bake the center of one particular named `Circle`, or
```
forall Circle C {
bake C.icon.center.x
bake C.icon.center.y
}
```
to bake the centers of all circles.
| 2.0 | feat: Make it possible to "bake" diagrams - ### Issue
Penrose uses random sampling to initialize variables (e.g., those marked `?` in a Style program, or shape properties that are not explicitly specified). If a user is happy with a particular instance of a diagram, they would like to be able to reproduce this diagram later down the road. For this reason, we introduced at some point the notion of a _variation_, i.e., a string that is used to determine the random seed prior to diagram generation. However, variations do not completely solve the problem of reproducibility, since changes to the system itself (e.g,. the compiler or optimizer) can also change the appearance of the final diagram. Hence, it would be nice to have a more "iron clad" solution for finalizing diagrams.
One answer is to simply say: the final diagram can be stored as an SVG. However, saving as SVG does not preserve editability of non-optimized features of the diagrams (e.g., colors, line widths, text labels, etc.).
### Proposal
Instead, this issue raises the following proposal:
> Add functionality to "bake" diagrams, by writing a Style file where all optimized attributes are explicitly overriden with the optimized constants.
For instance, suppose we have the programs
```
-- Domain
type Circle
```
```
-- Substance
Circle c1
```
```
-- Style
canvas {
width = 100
height = 100
}
forall Circle C {
shape C.icon = Circle {
r: 25
center: (?,?)
}
}
```
and the optimizer yields a center `(-20.21, 15.45)` for the circle. Then to bake this diagram, we would append the following lines to the Style program:
```
forall Circle `c1` {
override `c1`.icon.center.x = -20.21
override `c1`.icon.center.y = 15.45
}
```
This feature could be exposed by an additional IDE button like `Bake` that downloads the augmented Style, and a command line option `--bake-style` that likewise dumps the augmented Style.
### Implementation
There's probably some grungy stuff that would have to happen to implement this feature, but the core thing is just getting a list of (i) optimized variables, (ii) their paths in the Style program, and (iii) their final, optimized values, which shouldn't be too hard to get our hands on.
### Field-specific baking
An elaboration on this design would be to allow the Style file to specify properties that should be baked, using the usual selector paradigm. E.g., one could write
```
forall Circle `c1` {
bake `c1`.icon.center.x
bake `c1`.icon.center.y
}
```
to bake the center of one particular named `Circle`, or
```
forall Circle C {
bake C.icon.center.x
bake C.icon.center.y
}
```
to bake the centers of all circles.
| priority | feat make it possible to bake diagrams issue penrose uses random sampling to initialize variables e g those marked in a style program or shape properties that are not explicitly specified if a user is happy with a particular instance of a diagram they would like to be able to reproduce this diagram later down the road for this reason we introduced at some point the notion of a variation i e a string that is used to determine the random seed prior to diagram generation however variations do not completely solve the problem of reproducibility since changes to the system itself e g the compiler or optimizer can also change the appearance of the final diagram hence it would be nice to have a more iron clad solution for finalizing diagrams one answer is to simply say the final diagram can be stored as an svg however saving as svg does not preserve editability of non optimized features of the diagrams e g colors line widths text labels etc proposal instead this issue raises the following proposal add functionality to bake diagrams by writing a style file where all optimized attributes are explicitly overriden with the optimized constants for instance suppose we have the programs domain type circle substance circle style canvas width height forall circle c shape c icon circle r center and the optimizer yields a center for the circle then to bake this diagram we would append the following lines to the style program forall circle override icon center x override icon center y this feature could be exposed by an additional ide button like bake that downloads the augmented style and a command line option bake style that likewise dumps the augmented style implementation there s probably some grungy stuff that would have to happen to implement this feature but the core thing is just getting a list of i optimized variables ii their paths in the style program and iii their final optimized values which shouldn t be too hard to get our hands on field specific baking an elaboration on this design would be to allow the style file to specify properties that should be baked using the usual selector paradigm e g one could write forall circle bake icon center x bake icon center y to bake the center of one particular named circle or forall circle c bake c icon center x bake c icon center y to bake the centers of all circles | 1 |
25,553 | 2,683,840,932 | IssuesEvent | 2015-03-28 11:22:51 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | ConEmu 2009.10.1 - Не работает символ "c" | 2–5 stars bug imported Priority-Medium | _From [skonys...@gmail.com](https://code.google.com/u/113491849526375923289/) on October 13, 2009 10:18:24_
Версия ОС: WinXP Professional SP3
Не работает символ "c" в запущеном cmd.exe. При этом в фаре работает.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=112_ | 1.0 | ConEmu 2009.10.1 - Не работает символ "c" - _From [skonys...@gmail.com](https://code.google.com/u/113491849526375923289/) on October 13, 2009 10:18:24_
Версия ОС: WinXP Professional SP3
Не работает символ "c" в запущеном cmd.exe. При этом в фаре работает.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=112_ | priority | conemu не работает символ c from on october версия ос winxp professional не работает символ c в запущеном cmd exe при этом в фаре работает original issue | 1 |
118,518 | 25,324,196,225 | IssuesEvent | 2022-11-18 07:47:48 | haproxy/haproxy | https://api.github.com/repos/haproxy/haproxy | closed | src/mux_h1.c: null pointer dereference suspected by coverity | type: code-report | ### Tool Name and Version
coverity
### Code Report
```plain
*** CID 1501001: Null pointer dereferences (REVERSE_INULL)
/src/mux_h1.c: 3541 in h1_shutw_conn()
3535 TRACE_ENTER(H1_EV_H1C_END, conn);
3536 h1_close(h1c);
3537 if (conn->flags & CO_FL_SOCK_WR_SH)
3538 return;
3539
3540 conn_xprt_shutw(conn);
>>> CID 1501001: Null pointer dereferences (REVERSE_INULL)
>>> Null-checking "h1c" suggests that it may be null, but it has already been dereferenced on all paths leading to the check.
3541 conn_sock_shutw(conn, (h1c && !(h1c->flags & H1C_F_SILENT_SHUT)));
3542 TRACE_LEAVE(H1_EV_H1C_END, conn);
3543 }
3544
3545 /* Called from the upper layer, to unsubscribe <es> from events <event_type>
3546 * The <es> pointer is not allowed to differ from the one passed to the
```
### Additional Information
_No response_
### Output of `haproxy -vv`
```plain
no
```
| 1.0 | src/mux_h1.c: null pointer dereference suspected by coverity - ### Tool Name and Version
coverity
### Code Report
```plain
*** CID 1501001: Null pointer dereferences (REVERSE_INULL)
/src/mux_h1.c: 3541 in h1_shutw_conn()
3535 TRACE_ENTER(H1_EV_H1C_END, conn);
3536 h1_close(h1c);
3537 if (conn->flags & CO_FL_SOCK_WR_SH)
3538 return;
3539
3540 conn_xprt_shutw(conn);
>>> CID 1501001: Null pointer dereferences (REVERSE_INULL)
>>> Null-checking "h1c" suggests that it may be null, but it has already been dereferenced on all paths leading to the check.
3541 conn_sock_shutw(conn, (h1c && !(h1c->flags & H1C_F_SILENT_SHUT)));
3542 TRACE_LEAVE(H1_EV_H1C_END, conn);
3543 }
3544
3545 /* Called from the upper layer, to unsubscribe <es> from events <event_type>
3546 * The <es> pointer is not allowed to differ from the one passed to the
```
### Additional Information
_No response_
### Output of `haproxy -vv`
```plain
no
```
| non_priority | src mux c null pointer dereference suspected by coverity tool name and version coverity code report plain cid null pointer dereferences reverse inull src mux c in shutw conn trace enter ev end conn close if conn flags co fl sock wr sh return conn xprt shutw conn cid null pointer dereferences reverse inull null checking suggests that it may be null but it has already been dereferenced on all paths leading to the check conn sock shutw conn flags f silent shut trace leave ev end conn called from the upper layer to unsubscribe from events the pointer is not allowed to differ from the one passed to the additional information no response output of haproxy vv plain no | 0 |
332,878 | 10,112,261,065 | IssuesEvent | 2019-07-30 14:23:06 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | forums.warframe.com - design is broken | browser-firefox engine-gecko priority-normal status-needsinfo | <!-- @browser: Firefox 69.0b4 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: -->
**URL**: https://forums.warframe.com
**Browser / Version**: Firefox 69.0b4
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Picture Thumbnails aren't scaled down properly
**Steps to Reproduce**:
When you link pictures that exceed the width of the forum post reserved space, the image scales down.
As of the beta update from approximately 12/07/2019, the website has failed to scale down pictures appropiately and doesn't keep the aspect ratio. If I post a 1920x1080 picture, instead of scaling down to another 16:9 ratio image, it scales down the width but not the height. Console showcases width of 1200 when it's actually 1920 pixels.
Adding two imgur links since I can only attach a single screenshot:
Issue: https://i.imgur.com/P9HvpUn.png
Fixed manually changing value: https://i.imgur.com/Cg5orTG.png
Only happens on Firefox 69 beta, other users don't seem to see it and Chrome works fine. Please notice the Warframe Forums use InvisionBoard, so it's possible other sites might have similar issues.
[](https://webcompat.com/uploads/2019/7/b5e7d7ad-db08-4470-8d72-b6e4b0e041ed.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | forums.warframe.com - design is broken - <!-- @browser: Firefox 69.0b4 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: -->
**URL**: https://forums.warframe.com
**Browser / Version**: Firefox 69.0b4
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Picture Thumbnails aren't scaled down properly
**Steps to Reproduce**:
When you link pictures that exceed the width of the forum post reserved space, the image scales down.
As of the beta update from approximately 12/07/2019, the website has failed to scale down pictures appropiately and doesn't keep the aspect ratio. If I post a 1920x1080 picture, instead of scaling down to another 16:9 ratio image, it scales down the width but not the height. Console showcases width of 1200 when it's actually 1920 pixels.
Adding two imgur links since I can only attach a single screenshot:
Issue: https://i.imgur.com/P9HvpUn.png
Fixed manually changing value: https://i.imgur.com/Cg5orTG.png
Only happens on Firefox 69 beta, other users don't seem to see it and Chrome works fine. Please notice the Warframe Forums use InvisionBoard, so it's possible other sites might have similar issues.
[](https://webcompat.com/uploads/2019/7/b5e7d7ad-db08-4470-8d72-b6e4b0e041ed.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | forums warframe com design is broken url browser version firefox operating system windows tested another browser yes problem type design is broken description picture thumbnails aren t scaled down properly steps to reproduce when you link pictures that exceed the width of the forum post reserved space the image scales down as of the beta update from approximately the website has failed to scale down pictures appropiately and doesn t keep the aspect ratio if i post a picture instead of scaling down to another ratio image it scales down the width but not the height console showcases width of when it s actually pixels adding two imgur links since i can only attach a single screenshot issue fixed manually changing value only happens on firefox beta other users don t seem to see it and chrome works fine please notice the warframe forums use invisionboard so it s possible other sites might have similar issues browser configuration none from with ❤️ | 1 |
320,984 | 9,791,895,275 | IssuesEvent | 2019-06-10 16:02:57 | cloud-custodian/cloud-custodian | https://api.github.com/repos/cloud-custodian/cloud-custodian | opened | Azure - CosmosDB Action - Set Provisioned Throughput | cloud/azure priority/P3 | This requires design as it can be set at database/container levels and currently we only enumerate Servers (although databases coming soon).
The use case here is stopping the bleeding on over-provisioned instances, so a blunt instrument may be adequate. | 1.0 | Azure - CosmosDB Action - Set Provisioned Throughput - This requires design as it can be set at database/container levels and currently we only enumerate Servers (although databases coming soon).
The use case here is stopping the bleeding on over-provisioned instances, so a blunt instrument may be adequate. | priority | azure cosmosdb action set provisioned throughput this requires design as it can be set at database container levels and currently we only enumerate servers although databases coming soon the use case here is stopping the bleeding on over provisioned instances so a blunt instrument may be adequate | 1 |
553,872 | 16,384,354,961 | IssuesEvent | 2021-05-17 08:31:50 | MathiasReker/Delfinen | https://api.github.com/repos/MathiasReker/Delfinen | closed | Create Discipline | feature request high priority required | Feature Request
- [x] Create Discpline class
Includes:
Style
Distance
- [x] Enum for Style
Various swimming styles
- [x] Enum for Distance
various Distances in swimming
| 1.0 | Create Discipline - Feature Request
- [x] Create Discpline class
Includes:
Style
Distance
- [x] Enum for Style
Various swimming styles
- [x] Enum for Distance
various Distances in swimming
| priority | create discipline feature request create discpline class includes style distance enum for style various swimming styles enum for distance various distances in swimming | 1 |
166,006 | 20,711,370,109 | IssuesEvent | 2022-03-12 01:12:51 | snowflakedb/snowflake-jdbc | https://api.github.com/repos/snowflakedb/snowflake-jdbc | opened | CVE-2020-36188 (High) detected in jackson-databind-2.9.8.jar | security vulnerability | ## CVE-2020-36188 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p>
<p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36188","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36188 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-36188 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p>
<p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36188","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws ua kqqkcs archiveextraction belggc fuidan snowflake jdbc depth dependencies arrow vector meta inf maven org apache arrow arrow vector pom xml path to vulnerable library sitory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource vulnerabilityurl | 0 |
36,302 | 5,042,512,149 | IssuesEvent | 2016-12-19 14:13:57 | halestudio/hale | https://api.github.com/repos/halestudio/hale | closed | Add SKOS Codelist reader | feature hale-support io prio-1-must to be tested | Allow to import codelists using the SKOS format from Files or URLs, and allow to assign these code lists to schema elements.
| 1.0 | Add SKOS Codelist reader - Allow to import codelists using the SKOS format from Files or URLs, and allow to assign these code lists to schema elements.
| non_priority | add skos codelist reader allow to import codelists using the skos format from files or urls and allow to assign these code lists to schema elements | 0 |
78,014 | 3,508,687,744 | IssuesEvent | 2016-01-08 19:02:54 | LunaNode/lobster | https://api.github.com/repos/LunaNode/lobster | closed | SolusVM with OpenVZ: burst memory on creation | low-priority | Context: for SolusVM we only require setup of a single Lobster plan. Then we use customdisk, customcpu, etc. to configure the plan so that it matches the one in the Lobster database.
For OpenVZ, the vserver-create returns an invalid custom burst memory error if we set custommemory. Currently we are using a work-around where we change the memory after provisioning instead of initially. We should improve on this if possible. | 1.0 | SolusVM with OpenVZ: burst memory on creation - Context: for SolusVM we only require setup of a single Lobster plan. Then we use customdisk, customcpu, etc. to configure the plan so that it matches the one in the Lobster database.
For OpenVZ, the vserver-create returns an invalid custom burst memory error if we set custommemory. Currently we are using a work-around where we change the memory after provisioning instead of initially. We should improve on this if possible. | priority | solusvm with openvz burst memory on creation context for solusvm we only require setup of a single lobster plan then we use customdisk customcpu etc to configure the plan so that it matches the one in the lobster database for openvz the vserver create returns an invalid custom burst memory error if we set custommemory currently we are using a work around where we change the memory after provisioning instead of initially we should improve on this if possible | 1 |
159,825 | 6,062,201,755 | IssuesEvent | 2017-06-14 08:50:56 | python/mypy | https://api.github.com/repos/python/mypy | opened | Use plugin to support SQLAlchemy table definitions | feature priority-2-low topic-plugins | We'll also need stubs for SQLAlchemy as they were removed from typeshed some time ago. Current work on stubs is happening at https://github.com/JelleZijlstra/sqlalchemy-stubs. | 1.0 | Use plugin to support SQLAlchemy table definitions - We'll also need stubs for SQLAlchemy as they were removed from typeshed some time ago. Current work on stubs is happening at https://github.com/JelleZijlstra/sqlalchemy-stubs. | priority | use plugin to support sqlalchemy table definitions we ll also need stubs for sqlalchemy as they were removed from typeshed some time ago current work on stubs is happening at | 1 |
223,630 | 7,459,317,228 | IssuesEvent | 2018-03-30 14:49:38 | GoogleCloudPlatform/google-cloud-dotnet | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-dotnet | reopened | Feature Request: Warn the user if they call app.UseGoogleTrace() after app.UseMvc() | api: trace priority: p2 | If they call `app.UseGoogleTrace()` after `app.UseMvc()` they'll see no errors and no warnings and no traces, and have no clue why.
If there is a way to detect and warn users in this scenario, that would save them a lot of time. | 1.0 | Feature Request: Warn the user if they call app.UseGoogleTrace() after app.UseMvc() - If they call `app.UseGoogleTrace()` after `app.UseMvc()` they'll see no errors and no warnings and no traces, and have no clue why.
If there is a way to detect and warn users in this scenario, that would save them a lot of time. | priority | feature request warn the user if they call app usegoogletrace after app usemvc if they call app usegoogletrace after app usemvc they ll see no errors and no warnings and no traces and have no clue why if there is a way to detect and warn users in this scenario that would save them a lot of time | 1 |
32,583 | 15,444,566,746 | IssuesEvent | 2021-03-08 10:34:22 | odpi/egeria-connector-crux | https://api.github.com/repos/odpi/egeria-connector-crux | opened | Improve findXByProperty searches with no type limiters | performance | Currently the `findEntitiesByProperty` and `findRelationshipsByProperty` methods are consistently timing out at larger scale factors of the CTS (40+) when no type limitations have been provided. Curiously, the same does not seem to occur with `findEntitiesByPropertyValue` and `findRelationshipsByPropertyValue`.
This is therefore most likely caused not by the Lucene search and reverse-lookup itself, but rather the (potentially large) OR-based set of conditions that are currently generated for the `...ByProperty` methods: specifically, when no type limiter is provided, the query generation logic looks for every valid combination of type and property and the generated query could contain 10-30+ such OR conditions -- akin to running 10-30+ separate queries and combining all of the results.
Given that the same problem does not occur with the `...ByPropertyValue` methods, it would likely be better remove these OR-based conditions (for any String parameter) and instead do something along these lines:
- Wildcard lookup against the Lucene index, and destructure both the document and matched attribute
- Validate that the matched attribute's name ends with the property passed to the `...ByProperty` method (probably best to check for `.<property_name>`)
- Only where these match return that particular result
This then bypasses the need for any OR-based logic at all (at least for String-based parameters). | True | Improve findXByProperty searches with no type limiters - Currently the `findEntitiesByProperty` and `findRelationshipsByProperty` methods are consistently timing out at larger scale factors of the CTS (40+) when no type limitations have been provided. Curiously, the same does not seem to occur with `findEntitiesByPropertyValue` and `findRelationshipsByPropertyValue`.
This is therefore most likely caused not by the Lucene search and reverse-lookup itself, but rather the (potentially large) OR-based set of conditions that are currently generated for the `...ByProperty` methods: specifically, when no type limiter is provided, the query generation logic looks for every valid combination of type and property and the generated query could contain 10-30+ such OR conditions -- akin to running 10-30+ separate queries and combining all of the results.
Given that the same problem does not occur with the `...ByPropertyValue` methods, it would likely be better remove these OR-based conditions (for any String parameter) and instead do something along these lines:
- Wildcard lookup against the Lucene index, and destructure both the document and matched attribute
- Validate that the matched attribute's name ends with the property passed to the `...ByProperty` method (probably best to check for `.<property_name>`)
- Only where these match return that particular result
This then bypasses the need for any OR-based logic at all (at least for String-based parameters). | non_priority | improve findxbyproperty searches with no type limiters currently the findentitiesbyproperty and findrelationshipsbyproperty methods are consistently timing out at larger scale factors of the cts when no type limitations have been provided curiously the same does not seem to occur with findentitiesbypropertyvalue and findrelationshipsbypropertyvalue this is therefore most likely caused not by the lucene search and reverse lookup itself but rather the potentially large or based set of conditions that are currently generated for the byproperty methods specifically when no type limiter is provided the query generation logic looks for every valid combination of type and property and the generated query could contain such or conditions akin to running separate queries and combining all of the results given that the same problem does not occur with the bypropertyvalue methods it would likely be better remove these or based conditions for any string parameter and instead do something along these lines wildcard lookup against the lucene index and destructure both the document and matched attribute validate that the matched attribute s name ends with the property passed to the byproperty method probably best to check for only where these match return that particular result this then bypasses the need for any or based logic at all at least for string based parameters | 0 |
383,697 | 26,561,800,517 | IssuesEvent | 2023-01-20 16:27:39 | GCTC-NTGC/gc-digital-talent | https://api.github.com/repos/GCTC-NTGC/gc-digital-talent | closed | Cookies - What they are and what's spawning them | documentation | ## Acceptance Criteria
- [ ] List all the cookies on prod
- [ ] Document: Source, purpose, do we control it. For all cookies
- [ ] Add the table or document with this info to our Repo Readme | 1.0 | Cookies - What they are and what's spawning them - ## Acceptance Criteria
- [ ] List all the cookies on prod
- [ ] Document: Source, purpose, do we control it. For all cookies
- [ ] Add the table or document with this info to our Repo Readme | non_priority | cookies what they are and what s spawning them acceptance criteria list all the cookies on prod document source purpose do we control it for all cookies add the table or document with this info to our repo readme | 0 |
38,255 | 8,444,015,750 | IssuesEvent | 2018-10-18 17:10:49 | MicrosoftDocs/live-share | https://api.github.com/repos/MicrosoftDocs/live-share | closed | [VS Code] Collaboration session deleting of all data suddenly. | area: co-edit bug vscode | **Good** Morning Microsoft,
Today I was with my friend using the Live-Share feature but suddenly Visual Studio Code has deleted each and every single data from the file we were working on and it stopped all VSC Live Share invites as it has stated the session has been stopped by the owner but it wasn't. Therefore this is a problem which I think you might need to fix. As well as this I was with my other friend yesterday editing a project when suddenly the Editor was not allowing us to type in or delete anything in the project. I hope you could fix this ASAP. Thank you Deniernal354 | 1.0 | [VS Code] Collaboration session deleting of all data suddenly. - **Good** Morning Microsoft,
Today I was with my friend using the Live-Share feature but suddenly Visual Studio Code has deleted each and every single data from the file we were working on and it stopped all VSC Live Share invites as it has stated the session has been stopped by the owner but it wasn't. Therefore this is a problem which I think you might need to fix. As well as this I was with my other friend yesterday editing a project when suddenly the Editor was not allowing us to type in or delete anything in the project. I hope you could fix this ASAP. Thank you Deniernal354 | non_priority | collaboration session deleting of all data suddenly good morning microsoft today i was with my friend using the live share feature but suddenly visual studio code has deleted each and every single data from the file we were working on and it stopped all vsc live share invites as it has stated the session has been stopped by the owner but it wasn t therefore this is a problem which i think you might need to fix as well as this i was with my other friend yesterday editing a project when suddenly the editor was not allowing us to type in or delete anything in the project i hope you could fix this asap thank you | 0 |
63,892 | 26,543,271,922 | IssuesEvent | 2023-01-19 21:14:21 | Azure/azure-powershell | https://api.github.com/repos/Azure/azure-powershell | closed | [Feature]: Request that 'Null' be added as an accepted value for the -OsType parameter for New-AzDiskUpdateConfig | Compute feature-request Service Attention customer-reported | ### Description of the new feature
I have hundreds of data disks that have erroneously had the OsType parameter set. This causes Azure to display metrics for the OS rather than the data disk.
I can clear the OsType parameter by submitting the following:
$diskupdateconfig = New-AzDiskUpdateConfig
But I cannot apply the new setting because Null is not an accepted value:
Update-AzDisk -ResourceGroupName 'AZG-AZ-BRIAN-RG' -DiskName 'TST1-Mike_datadisk_1' -DiskUpdate $diskupdateconfig
### Proposed implementation details (optional)
_No response_ | 1.0 | [Feature]: Request that 'Null' be added as an accepted value for the -OsType parameter for New-AzDiskUpdateConfig - ### Description of the new feature
I have hundreds of data disks that have erroneously had the OsType parameter set. This causes Azure to display metrics for the OS rather than the data disk.
I can clear the OsType parameter by submitting the following:
$diskupdateconfig = New-AzDiskUpdateConfig
But I cannot apply the new setting because Null is not an accepted value:
Update-AzDisk -ResourceGroupName 'AZG-AZ-BRIAN-RG' -DiskName 'TST1-Mike_datadisk_1' -DiskUpdate $diskupdateconfig
### Proposed implementation details (optional)
_No response_ | non_priority | request that null be added as an accepted value for the ostype parameter for new azdiskupdateconfig description of the new feature i have hundreds of data disks that have erroneously had the ostype parameter set this causes azure to display metrics for the os rather than the data disk i can clear the ostype parameter by submitting the following diskupdateconfig new azdiskupdateconfig but i cannot apply the new setting because null is not an accepted value update azdisk resourcegroupname azg az brian rg diskname mike datadisk diskupdate diskupdateconfig proposed implementation details optional no response | 0 |
777,495 | 27,281,970,566 | IssuesEvent | 2023-02-23 10:43:19 | uhh-cms/columnflow | https://api.github.com/repos/uhh-cms/columnflow | opened | Add tasks and helpers to write pyhf workspaces | enhancement low-priority | We define our statistical models in an experiment agnostic way, allowing for various formats for exporting actual fit models (stat. model + data). So far, we have a combine datacard writer and an accompanying task that creates cards.
We should provide the same mechanism for Pyhf workspaces, which should be fairly easy to achieve, but could help us leveraging bunch of benefits. | 1.0 | Add tasks and helpers to write pyhf workspaces - We define our statistical models in an experiment agnostic way, allowing for various formats for exporting actual fit models (stat. model + data). So far, we have a combine datacard writer and an accompanying task that creates cards.
We should provide the same mechanism for Pyhf workspaces, which should be fairly easy to achieve, but could help us leveraging bunch of benefits. | priority | add tasks and helpers to write pyhf workspaces we define our statistical models in an experiment agnostic way allowing for various formats for exporting actual fit models stat model data so far we have a combine datacard writer and an accompanying task that creates cards we should provide the same mechanism for pyhf workspaces which should be fairly easy to achieve but could help us leveraging bunch of benefits | 1 |
335,299 | 10,151,579,458 | IssuesEvent | 2019-08-05 20:42:56 | rust-windowing/winit | https://api.github.com/repos/rust-windowing/winit | opened | Implement proper grabbing behavior on macOS | Good first issue difficulty: average platform: macOS priority: normal type: platform parity | Right now on macOS, the cursor just gets locked to its current location, instead of getting locked to the area of the window. A proper solution that locks the cursor to the window exists and is documented, but hasn't been implemented yet.
See this comment:
https://github.com/rust-windowing/winit/blob/2442305bb7f5e0e9c95733536e5153129dde59db/src/platform_impl/macos/window.rs#L519-L523 | 1.0 | Implement proper grabbing behavior on macOS - Right now on macOS, the cursor just gets locked to its current location, instead of getting locked to the area of the window. A proper solution that locks the cursor to the window exists and is documented, but hasn't been implemented yet.
See this comment:
https://github.com/rust-windowing/winit/blob/2442305bb7f5e0e9c95733536e5153129dde59db/src/platform_impl/macos/window.rs#L519-L523 | priority | implement proper grabbing behavior on macos right now on macos the cursor just gets locked to its current location instead of getting locked to the area of the window a proper solution that locks the cursor to the window exists and is documented but hasn t been implemented yet see this comment | 1 |
702,357 | 24,120,928,985 | IssuesEvent | 2022-09-20 18:38:12 | buxx/rolling | https://api.github.com/repos/buxx/rolling | closed | Plantage sur l'action prendre | bug priority | > Après avoir fini la construction du four en brique crue j ai voulu récupérer les briques surnuméraires dans le four. J ai cliqué sur prendre puis brique crue, le menu a disparu et le jeu a freezé totalement. Reproduit 2 fois après relance du jeu | 1.0 | Plantage sur l'action prendre - > Après avoir fini la construction du four en brique crue j ai voulu récupérer les briques surnuméraires dans le four. J ai cliqué sur prendre puis brique crue, le menu a disparu et le jeu a freezé totalement. Reproduit 2 fois après relance du jeu | priority | plantage sur l action prendre après avoir fini la construction du four en brique crue j ai voulu récupérer les briques surnuméraires dans le four j ai cliqué sur prendre puis brique crue le menu a disparu et le jeu a freezé totalement reproduit fois après relance du jeu | 1 |
428,862 | 12,418,334,163 | IssuesEvent | 2020-05-22 23:49:52 | WordPress/wordcamp.org | https://api.github.com/repos/WordPress/wordcamp.org | closed | Wildcard Certificate is Expired (since April 2018) | [Component] HTTPS [Priority] 4 [Status] Blocked [Type] Enhancement | **Describe the bug**
An expired wildcard certificate is being offered for "unknown" domains.
**To reproduce**
Steps to reproduce the behavior:
1. Go to https://example.wordcamp.org/
2. See browser error
**Expected behavior**
I would expect a friendly error message from wordcamp.org
| 1.0 | Wildcard Certificate is Expired (since April 2018) - **Describe the bug**
An expired wildcard certificate is being offered for "unknown" domains.
**To reproduce**
Steps to reproduce the behavior:
1. Go to https://example.wordcamp.org/
2. See browser error
**Expected behavior**
I would expect a friendly error message from wordcamp.org
| priority | wildcard certificate is expired since april describe the bug an expired wildcard certificate is being offered for unknown domains to reproduce steps to reproduce the behavior go to see browser error expected behavior i would expect a friendly error message from wordcamp org | 1 |
19,423 | 5,872,519,229 | IssuesEvent | 2017-05-15 11:46:07 | numbbo/coco | https://api.github.com/repos/numbbo/coco | closed | Merging run-LANGUAGE with install-postprocessing | Code-Experiments Documentation enhancement Priority-Medium Usability | In order to simplify the installation process, we should consider merging the `install-LANGUAGE` and the `install-postprocessing` parts of `do.py` to one command per language and (re-)name it to something like `install-and-test-LANGUAGE` or `install-coco-in-LANGUAGE`. The documentation should of course be updated accordingly.
| 1.0 | Merging run-LANGUAGE with install-postprocessing - In order to simplify the installation process, we should consider merging the `install-LANGUAGE` and the `install-postprocessing` parts of `do.py` to one command per language and (re-)name it to something like `install-and-test-LANGUAGE` or `install-coco-in-LANGUAGE`. The documentation should of course be updated accordingly.
| non_priority | merging run language with install postprocessing in order to simplify the installation process we should consider merging the install language and the install postprocessing parts of do py to one command per language and re name it to something like install and test language or install coco in language the documentation should of course be updated accordingly | 0 |
61,170 | 12,150,646,740 | IssuesEvent | 2020-04-24 18:22:19 | nopSolutions/nopCommerce | https://api.github.com/repos/nopSolutions/nopCommerce | closed | SQL Server password should be hidden during installation | refactoring / source code | nopCommerce version: 4.30 beta
Steps to reproduce the problem:
Try to install the nopcommerce 4.30 beta ,
At installation page in SQL Password, text box type is change password to text
Here is the screenshot: https://prnt.sc/rsjawo
| 1.0 | SQL Server password should be hidden during installation - nopCommerce version: 4.30 beta
Steps to reproduce the problem:
Try to install the nopcommerce 4.30 beta ,
At installation page in SQL Password, text box type is change password to text
Here is the screenshot: https://prnt.sc/rsjawo
| non_priority | sql server password should be hidden during installation nopcommerce version beta steps to reproduce the problem try to install the nopcommerce beta at installation page in sql password text box type is change password to text here is the screenshot | 0 |
245,641 | 7,889,028,983 | IssuesEvent | 2018-06-28 01:29:14 | Kheiden/RPi-tankbot | https://api.github.com/repos/Kheiden/RPi-tankbot | closed | Calibrate stereo cameras | Priority 2 core | Depends on #42
I want to remove distortion on the stereo cameras by calibrating them. I'll save the calibration for re-use. | 1.0 | Calibrate stereo cameras - Depends on #42
I want to remove distortion on the stereo cameras by calibrating them. I'll save the calibration for re-use. | priority | calibrate stereo cameras depends on i want to remove distortion on the stereo cameras by calibrating them i ll save the calibration for re use | 1 |
304,349 | 26,270,409,485 | IssuesEvent | 2023-01-06 16:26:51 | skeletonlabs/skeleton | https://api.github.com/repos/skeletonlabs/skeleton | closed | Update Plugin to support Tailwind Intellisense | enhancement ready to test | ### Current Behavior
I've heard reports it works for Daisy UI, and am sure it works locally in the Skeleton project, so we'll need to see how to carry this through to end users.
### Steps To Reproduce
_No response_
### Anything else?
_No response_ | 1.0 | Update Plugin to support Tailwind Intellisense - ### Current Behavior
I've heard reports it works for Daisy UI, and am sure it works locally in the Skeleton project, so we'll need to see how to carry this through to end users.
### Steps To Reproduce
_No response_
### Anything else?
_No response_ | non_priority | update plugin to support tailwind intellisense current behavior i ve heard reports it works for daisy ui and am sure it works locally in the skeleton project so we ll need to see how to carry this through to end users steps to reproduce no response anything else no response | 0 |
557,752 | 16,517,983,070 | IssuesEvent | 2021-05-26 11:48:15 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | When uploading photos and viewing them in mobile screen mode, some of the photos will overlap and some will not display in the activity feed | bug priority: medium | **Describe the bug**
When uploading photos and viewing them in mobile screen mode, some of the photos will overlap and some will not display in the activity feed.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Activity feed'
2. Click on 'upload photo'
3. 'When uploading photos and viewing them in mobile screen mode
4. some of the photos will overlap and some will not display in the activity feed.
**Expected behavior**
It should not overlap and some will display in the activity feed.
**Screenshots**
https://prnt.sc/13fsk7z
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| 1.0 | When uploading photos and viewing them in mobile screen mode, some of the photos will overlap and some will not display in the activity feed - **Describe the bug**
When uploading photos and viewing them in mobile screen mode, some of the photos will overlap and some will not display in the activity feed.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Activity feed'
2. Click on 'upload photo'
3. 'When uploading photos and viewing them in mobile screen mode
4. some of the photos will overlap and some will not display in the activity feed.
**Expected behavior**
It should not overlap and some will display in the activity feed.
**Screenshots**
https://prnt.sc/13fsk7z
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| priority | when uploading photos and viewing them in mobile screen mode some of the photos will overlap and some will not display in the activity feed describe the bug when uploading photos and viewing them in mobile screen mode some of the photos will overlap and some will not display in the activity feed to reproduce steps to reproduce the behavior go to activity feed click on upload photo when uploading photos and viewing them in mobile screen mode some of the photos will overlap and some will not display in the activity feed expected behavior it should not overlap and some will display in the activity feed screenshots support ticket links if applicable add helpscout link or ticket number where the issue was originally reported | 1 |
683,644 | 23,389,825,015 | IssuesEvent | 2022-08-11 16:42:39 | pixie-io/pixie | https://api.github.com/repos/pixie-io/pixie | closed | `px live` ctrl+k scripts aren't sorted | priority/backlog triage/accepted area/cli | We would expect the scripts shown in `ctrl + k` to be sorted in some reasonable order, so that scripts are easier for the user to filter through.
!image-2021-10-08-14-04-29-654.png|width=200,height=183!
┆Issue is synchronized with this [Jira Bug](https://pixie-labs.atlassian.net/browse/PX-132) by [Unito](https://www.unito.io)
┆Attachments: <a href="https://pixie-labs.atlassian.net/rest/api/2/attachment/content/10621">image-2021-10-08-14-04-29-654.png</a>
| 1.0 | `px live` ctrl+k scripts aren't sorted - We would expect the scripts shown in `ctrl + k` to be sorted in some reasonable order, so that scripts are easier for the user to filter through.
!image-2021-10-08-14-04-29-654.png|width=200,height=183!
┆Issue is synchronized with this [Jira Bug](https://pixie-labs.atlassian.net/browse/PX-132) by [Unito](https://www.unito.io)
┆Attachments: <a href="https://pixie-labs.atlassian.net/rest/api/2/attachment/content/10621">image-2021-10-08-14-04-29-654.png</a>
| priority | px live ctrl k scripts aren t sorted we would expect the scripts shown in ctrl k to be sorted in some reasonable order so that scripts are easier for the user to filter through image png width height ┆issue is synchronized with this by ┆attachments a href | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.