Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
212,693 | 16,473,470,212 | IssuesEvent | 2021-05-23 21:48:20 | reclosedev/requests-cache | https://api.github.com/repos/reclosedev/requests-cache | closed | Run tests against multiple versions of requests library | tests | It would be worthwhile to run tests against multiple versions of `requests`, let's say the previous 4 (or more?) minor versions in addition to the latest version. Currently this would be `2.25.1` (latest), `2.24.0`, `2.23.0`, `2.22.0`, `2.21.0`.
This should be relatively simple using a version matrix in GitHub Actions. Since this would add some time to CI jobs (even if run in parallel), maybe this should only be run for merges into `master` and skipped for other builds like pull requests. | 1.0 | Run tests against multiple versions of requests library - It would be worthwhile to run tests against multiple versions of `requests`, let's say the previous 4 (or more?) minor versions in addition to the latest version. Currently this would be `2.25.1` (latest), `2.24.0`, `2.23.0`, `2.22.0`, `2.21.0`.
This should be relatively simple using a version matrix in GitHub Actions. Since this would add some time to CI jobs (even if run in parallel), maybe this should only be run for merges into `master` and skipped for other builds like pull requests. | non_process | run tests against multiple versions of requests library it would be worthwhile to run tests against multiple versions of requests let s say the previous or more minor versions in addition to the latest version currently this would be latest this should be relatively simple using a version matrix in github actions since this would add some time to ci jobs even if run in parallel maybe this should only be run for merges into master and skipped for other builds like pull requests | 0 |
115,019 | 9,778,827,319 | IssuesEvent | 2019-06-07 13:06:16 | ValveSoftware/halflife | https://api.github.com/repos/ValveSoftware/halflife | closed | [CS 1.6][REQ] Music in MOTD | Need Retest reviewed | Another anti-annoying-music request. Thank you for adding _speak_enable_ cvar, it works perfectly, but there's an another way to play high volume music on 1.6 server.
With [this code](http://pastebin.com/Uhq2tvQm) in MOTD all players will listen for music if they have **volume 0**, **MP3Volume 0** and even after they leave your server. Song will play until it ends or till you quit cs.
Windows 7 x86, build 5958.
| 1.0 | [CS 1.6][REQ] Music in MOTD - Another anti-annoying-music request. Thank you for adding _speak_enable_ cvar, it works perfectly, but there's an another way to play high volume music on 1.6 server.
With [this code](http://pastebin.com/Uhq2tvQm) in MOTD all players will listen for music if they have **volume 0**, **MP3Volume 0** and even after they leave your server. Song will play until it ends or till you quit cs.
Windows 7 x86, build 5958.
| non_process | music in motd another anti annoying music request thank you for adding speak enable cvar it works perfectly but there s an another way to play high volume music on server with in motd all players will listen for music if they have volume and even after they leave your server song will play until it ends or till you quit cs windows build | 0 |
32,609 | 12,133,323,307 | IssuesEvent | 2020-04-23 08:50:45 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | JIT policy is not clear on usage | Pri1 assigned-to-author doc-enhancement security-center/svc triaged | When creating a policy that controls access to RDP port 3389 for the CIDR block of 10.21.254.0/24 I would expect that anyone making a JIT access request from within this range would be able to do so, e.g requesting connections from 10.21.254.85 to be allowed. This, however, is not the case, the only allowed connection from my experiments is when you request for the entire block to allowed all at once, 10.21.254.0/24.
This to me doesn't really make a lot of sense, and I'm pretty sure this is not the intention of JIT or am i mistaken and the JIT policy has to be matched verbatim?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 25f0758b-8274-d2b5-4702-2c7ca6b0b14a
* Version Independent ID: d1215107-abf9-fefd-a63b-b20ec3f7f744
* Content: [Just-in-time virtual machine access in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-just-in-time#feedback)
* Content Source: [articles/security-center/security-center-just-in-time.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-just-in-time.md)
* Service: **security-center**
* GitHub Login: @memildin
* Microsoft Alias: **memildin** | True | JIT policy is not clear on usage - When creating a policy that controls access to RDP port 3389 for the CIDR block of 10.21.254.0/24 I would expect that anyone making a JIT access request from within this range would be able to do so, e.g requesting connections from 10.21.254.85 to be allowed. This, however, is not the case, the only allowed connection from my experiments is when you request for the entire block to allowed all at once, 10.21.254.0/24.
This to me doesn't really make a lot of sense, and I'm pretty sure this is not the intention of JIT or am i mistaken and the JIT policy has to be matched verbatim?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 25f0758b-8274-d2b5-4702-2c7ca6b0b14a
* Version Independent ID: d1215107-abf9-fefd-a63b-b20ec3f7f744
* Content: [Just-in-time virtual machine access in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-just-in-time#feedback)
* Content Source: [articles/security-center/security-center-just-in-time.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-just-in-time.md)
* Service: **security-center**
* GitHub Login: @memildin
* Microsoft Alias: **memildin** | non_process | jit policy is not clear on usage when creating a policy that controls access to rdp port for the cidr block of i would expect that anyone making a jit access request from within this range would be able to do so e g requesting connections from to be allowed this however is not the case the only allowed connection from my experiments is when you request for the entire block to allowed all at once this to me doesn t really make a lot of sense and i m pretty sure this is not the intention of jit or am i mistaken and the jit policy has to be matched verbatim document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id fefd content content source service security center github login memildin microsoft alias memildin | 0 |
56,183 | 11,525,125,113 | IssuesEvent | 2020-02-15 05:51:13 | jsbxyyx/x-comment | https://api.github.com/repos/jsbxyyx/x-comment | opened | leetcode 0005 最长回文子串 | xxzkid | 逗的8笔 | Gitalk leetcode-0005 | https://uyiplus.com/2020/leetcode-0005/
给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 1000。 示例 1:123输入: "babad"输出: "bab"注意: "aba" 也是一个有效答案。 示例 2:12输入: "cbbd"输出: "bb" | 1.0 | leetcode 0005 最长回文子串 | xxzkid | 逗的8笔 - https://uyiplus.com/2020/leetcode-0005/
给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 1000。 示例 1:123输入: "babad"输出: "bab"注意: "aba" 也是一个有效答案。 示例 2:12输入: "cbbd"输出: "bb" | non_process | leetcode 最长回文子串 xxzkid 给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 。 示例 : quot babad quot 输出 quot bab quot 注意 quot aba quot 也是一个有效答案。 示例 : quot cbbd quot 输出 quot bb quot | 0 |
184,261 | 14,284,063,288 | IssuesEvent | 2020-11-23 11:57:19 | rtCamp/rtMedia | https://api.github.com/repos/rtCamp/rtMedia | closed | [Dashbaord] Hire Us tab - Correct the back ground color of the button | Low Tested UI | It should be `#007cba` color.

| 1.0 | [Dashbaord] Hire Us tab - Correct the back ground color of the button - It should be `#007cba` color.

| non_process | hire us tab correct the back ground color of the button it should be color | 0 |
32,121 | 6,717,569,809 | IssuesEvent | 2017-10-14 23:04:28 | k3dar/rdiff-backup-fs | https://api.github.com/repos/k3dar/rdiff-backup-fs | closed | Can't compile | auto-migrated Priority-Medium Type-Defect | _From @GoogleCodeExporter on July 5, 2015 0:23_
```
Hi, I'm using current debian testing (squeeze).
Here's what I get after ./configure while trying to make:
gcc -c -Wall `pkg-config --cflags fuse` -D_GNU_SOURCE archfs.c
/bin/sh: pkg-config: not found
In file included from /usr/include/fuse/fuse.h:26,
from /usr/include/fuse.h:9,
from headers.h:21,
from archfs.c:1:
/usr/include/fuse/fuse_common.h:32:2: error: #error Please add
-D_FILE_OFFSET_BITS=64 to your compile flags!
make: *** [archfs.o] Error 1
Configure output:
checking for gcc... gcc
checking whether the C compiler works... yes
checking for C compiler default output file name... a.out
checking for suffix of executables...
checking whether we are cross compiling... no
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for fuse_main in -lfuse... yes
checking how to run the C preprocessor... gcc -E
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking zlib.h usability... yes
checking zlib.h presence... yes
checking for zlib.h... yes
checking stdio.h usability... yes
checking stdio.h presence... yes
checking for stdio.h... yes
checking errno.h usability... yes
checking errno.h presence... yes
checking for errno.h... yes
checking stdarg.h usability... yes
checking stdarg.h presence... yes
checking for stdarg.h... yes
checking dirent.h usability... yes
checking dirent.h presence... yes
checking for dirent.h... yes
checking sys/wait.h usability... yes
checking sys/wait.h presence... yes
checking for sys/wait.h... yes
checking pthread.h usability... yes
checking pthread.h presence... yes
checking for pthread.h... yes
checking for get_current_dir_name... yes
checking for getline... yes
configure: creating ./config.status
```
Original issue reported on code.google.com by `snajdr.p...@gmail.com` on 29 Dec 2010 at 2:41
_Copied from original issue: rbrito/rdiff-backup-fs#1_ | 1.0 | Can't compile - _From @GoogleCodeExporter on July 5, 2015 0:23_
```
Hi, I'm using current debian testing (squeeze).
Here's what I get after ./configure while trying to make:
gcc -c -Wall `pkg-config --cflags fuse` -D_GNU_SOURCE archfs.c
/bin/sh: pkg-config: not found
In file included from /usr/include/fuse/fuse.h:26,
from /usr/include/fuse.h:9,
from headers.h:21,
from archfs.c:1:
/usr/include/fuse/fuse_common.h:32:2: error: #error Please add
-D_FILE_OFFSET_BITS=64 to your compile flags!
make: *** [archfs.o] Error 1
Configure output:
checking for gcc... gcc
checking whether the C compiler works... yes
checking for C compiler default output file name... a.out
checking for suffix of executables...
checking whether we are cross compiling... no
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for fuse_main in -lfuse... yes
checking how to run the C preprocessor... gcc -E
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking zlib.h usability... yes
checking zlib.h presence... yes
checking for zlib.h... yes
checking stdio.h usability... yes
checking stdio.h presence... yes
checking for stdio.h... yes
checking errno.h usability... yes
checking errno.h presence... yes
checking for errno.h... yes
checking stdarg.h usability... yes
checking stdarg.h presence... yes
checking for stdarg.h... yes
checking dirent.h usability... yes
checking dirent.h presence... yes
checking for dirent.h... yes
checking sys/wait.h usability... yes
checking sys/wait.h presence... yes
checking for sys/wait.h... yes
checking pthread.h usability... yes
checking pthread.h presence... yes
checking for pthread.h... yes
checking for get_current_dir_name... yes
checking for getline... yes
configure: creating ./config.status
```
Original issue reported on code.google.com by `snajdr.p...@gmail.com` on 29 Dec 2010 at 2:41
_Copied from original issue: rbrito/rdiff-backup-fs#1_ | non_process | can t compile from googlecodeexporter on july hi i m using current debian testing squeeze here s what i get after configure while trying to make gcc c wall pkg config cflags fuse d gnu source archfs c bin sh pkg config not found in file included from usr include fuse fuse h from usr include fuse h from headers h from archfs c usr include fuse fuse common h error error please add d file offset bits to your compile flags make error configure output checking for gcc gcc checking whether the c compiler works yes checking for c compiler default output file name a out checking for suffix of executables checking whether we are cross compiling no checking for suffix of object files o checking whether we are using the gnu c compiler yes checking whether gcc accepts g yes checking for gcc option to accept iso none needed checking for fuse main in lfuse yes checking how to run the c preprocessor gcc e checking for grep that handles long lines and e bin grep checking for egrep bin grep e checking for ansi c header files yes checking for sys types h yes checking for sys stat h yes checking for stdlib h yes checking for string h yes checking for memory h yes checking for strings h yes checking for inttypes h yes checking for stdint h yes checking for unistd h yes checking zlib h usability yes checking zlib h presence yes checking for zlib h yes checking stdio h usability yes checking stdio h presence yes checking for stdio h yes checking errno h usability yes checking errno h presence yes checking for errno h yes checking stdarg h usability yes checking stdarg h presence yes checking for stdarg h yes checking dirent h usability yes checking dirent h presence yes checking for dirent h yes checking sys wait h usability yes checking sys wait h presence yes checking for sys wait h yes checking pthread h usability yes checking pthread h presence yes checking for pthread h yes checking for get current dir name yes checking for getline yes configure creating config status original issue reported on code google com by snajdr p gmail com on dec at copied from original issue rbrito rdiff backup fs | 0 |
65,461 | 12,597,829,650 | IssuesEvent | 2020-06-11 01:04:45 | grpc/grpc-java | https://api.github.com/repos/grpc/grpc-java | opened | xds: investigate io.envoyproxy.controlplane:api | code health | Similar to #6016 which was not possible because the latest version of `io.envoyproxy.protoc-gen-validate:pgv-java-stub` in maven is not up to date as what we needed.
`io.envoyproxy.controlplane:api` in maven is more up to date and even includes `com.google.cloudapi` that #7109 needs. However, we don't know how frequently it updates on a regular basis. It may still fall behind our need. | 1.0 | xds: investigate io.envoyproxy.controlplane:api - Similar to #6016 which was not possible because the latest version of `io.envoyproxy.protoc-gen-validate:pgv-java-stub` in maven is not up to date as what we needed.
`io.envoyproxy.controlplane:api` in maven is more up to date and even includes `com.google.cloudapi` that #7109 needs. However, we don't know how frequently it updates on a regular basis. It may still fall behind our need. | non_process | xds investigate io envoyproxy controlplane api similar to which was not possible because the latest version of io envoyproxy protoc gen validate pgv java stub in maven is not up to date as what we needed io envoyproxy controlplane api in maven is more up to date and even includes com google cloudapi that needs however we don t know how frequently it updates on a regular basis it may still fall behind our need | 0 |
87,857 | 10,981,936,497 | IssuesEvent | 2019-12-01 02:41:43 | gopherjs/vecty | https://api.github.com/repos/gopherjs/vecty | closed | vecty.Render into non-body component | arch-design enhancement | Currently Render is restricted to a body component; i.e. Vecty must own the whole page effectively. We should allow rendering directly into an element that we own, e.g. `vecty.RenderInto("elem-id")` or something. | 1.0 | vecty.Render into non-body component - Currently Render is restricted to a body component; i.e. Vecty must own the whole page effectively. We should allow rendering directly into an element that we own, e.g. `vecty.RenderInto("elem-id")` or something. | non_process | vecty render into non body component currently render is restricted to a body component i e vecty must own the whole page effectively we should allow rendering directly into an element that we own e g vecty renderinto elem id or something | 0 |
804,730 | 29,498,635,953 | IssuesEvent | 2023-06-02 19:21:31 | asc-csa/Scisat-App | https://api.github.com/repos/asc-csa/Scisat-App | closed | Missing definition of error bars in altitude plot figure caption | science low priority | The figure caption should indicate the type of error bars.
With further edits of the text:
Graph [Plot] showing the gas concentration in parts per volume (ppv) over the selected altitude interval. The value represents the mean concentration over the selected latitudes, longitudes and dates. Error bars are 95% confidence intervals around the mean. | 1.0 | Missing definition of error bars in altitude plot figure caption - The figure caption should indicate the type of error bars.
With further edits of the text:
Graph [Plot] showing the gas concentration in parts per volume (ppv) over the selected altitude interval. The value represents the mean concentration over the selected latitudes, longitudes and dates. Error bars are 95% confidence intervals around the mean. | non_process | missing definition of error bars in altitude plot figure caption the figure caption should indicate the type of error bars with further edits of the text graph showing the gas concentration in parts per volume ppv over the selected altitude interval the value represents the mean concentration over the selected latitudes longitudes and dates error bars are confidence intervals around the mean | 0 |
15,137 | 11,371,924,594 | IssuesEvent | 2020-01-28 00:01:46 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Unpin Microsoft.Bcl.AsyncInterfaces dependency | area-infrastructure | We have [one pinned dependency on dotnet/corefx](https://github.com/dotnet/aspnetcore/blob/2cc333ac3013218a3985ecd0809b0c541cbe1f91/eng/Version.Details.xml#L405). It means our dependency flow widget has a big red box we can't get rid of :(

The dependency is `Microsoft.Bcl.AsyncInterfaces` (needed to build `IAsyncEnumerable` stuff on .NET Standard). It was pinned in https://github.com/dotnet/aspnetcore/pull/16716 by @BrennanConroy. @BrennanConroy do you know if we can unpin now? | 1.0 | Unpin Microsoft.Bcl.AsyncInterfaces dependency - We have [one pinned dependency on dotnet/corefx](https://github.com/dotnet/aspnetcore/blob/2cc333ac3013218a3985ecd0809b0c541cbe1f91/eng/Version.Details.xml#L405). It means our dependency flow widget has a big red box we can't get rid of :(

The dependency is `Microsoft.Bcl.AsyncInterfaces` (needed to build `IAsyncEnumerable` stuff on .NET Standard). It was pinned in https://github.com/dotnet/aspnetcore/pull/16716 by @BrennanConroy. @BrennanConroy do you know if we can unpin now? | non_process | unpin microsoft bcl asyncinterfaces dependency we have it means our dependency flow widget has a big red box we can t get rid of the dependency is microsoft bcl asyncinterfaces needed to build iasyncenumerable stuff on net standard it was pinned in by brennanconroy brennanconroy do you know if we can unpin now | 0 |
18,939 | 24,900,034,991 | IssuesEvent | 2022-10-28 19:48:05 | MicrosoftDocs/windows-dev-docs | https://api.github.com/repos/MicrosoftDocs/windows-dev-docs | closed | Hiding windowsupdate-optionalupdates hides all windows update panel not only optional updates | uwp/prod processes-and-threading/tech Pri1 | Hi,
If I want to hide the optional updates part from Windows Update panel, it will not only hide optional updates but all Windows Update panel.
The URI to use is the following one: windowsupdate-optionalupdates
I try with CSP and catalog settings as below:
hide:windowsupdate-optionalupdates
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 987ec16c-9456-93a4-177a-dbd563be7eb7
* Version Independent ID: f41f0344-f7f6-f092-a6bf-fc4184a9b460
* Content: [Launch the Windows Settings app - UWP applications](https://learn.microsoft.com/en-us/windows/uwp/launch-resume/launch-settings-app#ms-settings-uri-scheme-reference)
* Content Source: [windows-apps-src/launch-resume/launch-settings-app.md](https://github.com/MicrosoftDocs/windows-dev-docs/blob/docs/windows-apps-src/launch-resume/launch-settings-app.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft** | 1.0 | Hiding windowsupdate-optionalupdates hides all windows update panel not only optional updates - Hi,
If I want to hide the optional updates part from Windows Update panel, it will not only hide optional updates but all Windows Update panel.
The URI to use is the following one: windowsupdate-optionalupdates
I try with CSP and catalog settings as below:
hide:windowsupdate-optionalupdates
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 987ec16c-9456-93a4-177a-dbd563be7eb7
* Version Independent ID: f41f0344-f7f6-f092-a6bf-fc4184a9b460
* Content: [Launch the Windows Settings app - UWP applications](https://learn.microsoft.com/en-us/windows/uwp/launch-resume/launch-settings-app#ms-settings-uri-scheme-reference)
* Content Source: [windows-apps-src/launch-resume/launch-settings-app.md](https://github.com/MicrosoftDocs/windows-dev-docs/blob/docs/windows-apps-src/launch-resume/launch-settings-app.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft** | process | hiding windowsupdate optionalupdates hides all windows update panel not only optional updates hi if i want to hide the optional updates part from windows update panel it will not only hide optional updates but all windows update panel the uri to use is the following one windowsupdate optionalupdates i try with csp and catalog settings as below hide windowsupdate optionalupdates document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product uwp technology processes and threading github login alvinashcraft microsoft alias aashcraft | 1 |
3,197 | 4,236,926,162 | IssuesEvent | 2016-07-05 20:03:47 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | opened | Prepare System.Security.Cryptography.Native for OpenSSL 1.1 release | System.Security X-Plat | OpenSSL 1.1 has a new model of opaque structures with more accessor functions.
Anything previously obtained by an accessor function will likely be fine, but direct field accesses won't work.
For example, UpRefEvpPkey will need to change to EVP_PKEY_up_ref, but needs to do the current behavior on 1.0.x builds. | True | Prepare System.Security.Cryptography.Native for OpenSSL 1.1 release - OpenSSL 1.1 has a new model of opaque structures with more accessor functions.
Anything previously obtained by an accessor function will likely be fine, but direct field accesses won't work.
For example, UpRefEvpPkey will need to change to EVP_PKEY_up_ref, but needs to do the current behavior on 1.0.x builds. | non_process | prepare system security cryptography native for openssl release openssl has a new model of opaque structures with more accessor functions anything previously obtained by an accessor function will likely be fine but direct field accesses won t work for example uprefevppkey will need to change to evp pkey up ref but needs to do the current behavior on x builds | 0 |
241,526 | 26,256,819,694 | IssuesEvent | 2023-01-06 02:00:44 | benlazarine/cas-overlay | https://api.github.com/repos/benlazarine/cas-overlay | opened | WS-2019-0490 (High) detected in jcommander-1.48.jar | security vulnerability | ## WS-2019-0490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jcommander-1.48.jar</b></p></summary>
<p>A Java framework to parse command line options with annotations.</p>
<p>Library home page: <a href="http://beust.com/">http://beust.com/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/beust/jcommander/1.48/jcommander-1.48.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- pac4j-saml-3.4.0.jar
- xmlsectool-2.0.0.jar
- :x: **jcommander-1.48.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Inclusion of Functionality from Untrusted Control Sphere vulnerability found in jcommander before 1.75. jcommander resolving dependencies over HTTP instead of HTTPS.
<p>Publish Date: 2019-02-19
<p>URL: <a href=https://github.com/cbeust/jcommander/commit/3ae95595febbed9c13f367b6bda5c0be1c572c53>WS-2019-0490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-19</p>
<p>Fix Resolution (com.beust:jcommander): 1.75</p>
<p>Direct dependency fix Resolution (org.apereo.cas:cas-server-support-oauth-webflow): 6.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0490 (High) detected in jcommander-1.48.jar - ## WS-2019-0490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jcommander-1.48.jar</b></p></summary>
<p>A Java framework to parse command line options with annotations.</p>
<p>Library home page: <a href="http://beust.com/">http://beust.com/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/beust/jcommander/1.48/jcommander-1.48.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- pac4j-saml-3.4.0.jar
- xmlsectool-2.0.0.jar
- :x: **jcommander-1.48.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Inclusion of Functionality from Untrusted Control Sphere vulnerability found in jcommander before 1.75. jcommander resolving dependencies over HTTP instead of HTTPS.
<p>Publish Date: 2019-02-19
<p>URL: <a href=https://github.com/cbeust/jcommander/commit/3ae95595febbed9c13f367b6bda5c0be1c572c53>WS-2019-0490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-19</p>
<p>Fix Resolution (com.beust:jcommander): 1.75</p>
<p>Direct dependency fix Resolution (org.apereo.cas:cas-server-support-oauth-webflow): 6.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | ws high detected in jcommander jar ws high severity vulnerability vulnerable library jcommander jar a java framework to parse command line options with annotations library home page a href path to dependency file pom xml path to vulnerable library root repository com beust jcommander jcommander jar dependency hierarchy cas server support oauth webflow jar root library saml jar xmlsectool jar x jcommander jar vulnerable library vulnerability details inclusion of functionality from untrusted control sphere vulnerability found in jcommander before jcommander resolving dependencies over http instead of https publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com beust jcommander direct dependency fix resolution org apereo cas cas server support oauth webflow step up your open source security game with mend | 0 |
737,513 | 25,519,235,212 | IssuesEvent | 2022-11-28 18:58:25 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | closed | [opentitantool] Persistent sessions | Component:Software Component:Tooling Priority:P3 Type:Enhancement SW:opentitantool | # Problem
OpenTitan tool allows for a variety of backend "Transports", some of which are drivers for hardware debuggers, others emulate OpenTitan or related chips.
By their nature, some of those drivers will maintain state in the Rust struct that implements the Transport trait, others may have USB connections to hardware, which gets disturbed if the USB connection is closed and restablished.
Rust programs which link directly with OpenTitan lib are able to keep the Transport object for a long as needed, making a sequence of method invocations to "stimulate" the device under test. However, users of the command line OpenTitan tool does not have that ability, since each command line invocation generally results in the creation of a Transport object, and the one "useful" method invocation, before the Transport is closed as the process terminates.
# Proposal
We will add a new optional persistent “OpenTitan session proxy” process (a daemon). When used, all communication with the Transport backend will go through this proxy process. The proxy process will listen on a TCP socket and contain the "real" Transport implementation, maintaining emulator state or connection to USB hardware. The command line OpenTitan tool will use a "stub" Transport implementation, which forwards all requests via a TCP connection to the proxy.
Design document: https://docs.google.com/document/d/18JIqdYcmJKDzANheNS9JI1rEre6-iXPZhjMjliAS0TE/edit#
## Tasks
- [x] Make sure all error return values from Transport methods can be serialized (with serde) (#10964)
- [x] Declare wire protocol by as a number of Rust structs/enums (#10994)
- [x] Implement "stub" Transport (using serde JSON text via TCP) (#11023)
- [x] Implement proxy server functionality (opentitansession binary) (#11054)
- [x] Ability for opentitansession to run in background (as daemon) (#11173)
This proposal has significant overlap with previous issue #10217, which proposed a JSON TCP protocol for communicating with an emulator running in a separate process. With this proposal, the interface to the emulator would instead be the Transport trait itself, that is, the top-level process of the emulator would either be opentitansession, or would be another binary that links with a library containing most of the functionality of opentitansession (the listening on TCP socket, and translation of protocol into calls to Transport trait.) | 1.0 | [opentitantool] Persistent sessions - # Problem
OpenTitan tool allows for a variety of backend "Transports", some of which are drivers for hardware debuggers, others emulate OpenTitan or related chips.
By their nature, some of those drivers will maintain state in the Rust struct that implements the Transport trait, others may have USB connections to hardware, which gets disturbed if the USB connection is closed and restablished.
Rust programs which link directly with OpenTitan lib are able to keep the Transport object for a long as needed, making a sequence of method invocations to "stimulate" the device under test. However, users of the command line OpenTitan tool does not have that ability, since each command line invocation generally results in the creation of a Transport object, and the one "useful" method invocation, before the Transport is closed as the process terminates.
# Proposal
We will add a new optional persistent “OpenTitan session proxy” process (a daemon). When used, all communication with the Transport backend will go through this proxy process. The proxy process will listen on a TCP socket and contain the "real" Transport implementation, maintaining emulator state or connection to USB hardware. The command line OpenTitan tool will use a "stub" Transport implementation, which forwards all requests via a TCP connection to the proxy.
Design document: https://docs.google.com/document/d/18JIqdYcmJKDzANheNS9JI1rEre6-iXPZhjMjliAS0TE/edit#
## Tasks
- [x] Make sure all error return values from Transport methods can be serialized (with serde) (#10964)
- [x] Declare wire protocol by as a number of Rust structs/enums (#10994)
- [x] Implement "stub" Transport (using serde JSON text via TCP) (#11023)
- [x] Implement proxy server functionality (opentitansession binary) (#11054)
- [x] Ability for opentitansession to run in background (as daemon) (#11173)
This proposal has significant overlap with previous issue #10217, which proposed a JSON TCP protocol for communicating with an emulator running in a separate process. With this proposal, the interface to the emulator would instead be the Transport trait itself, that is, the top-level process of the emulator would either be opentitansession, or would be another binary that links with a library containing most of the functionality of opentitansession (the listening on TCP socket, and translation of protocol into calls to Transport trait.) | non_process | persistent sessions problem opentitan tool allows for a variety of backend transports some of which are drivers for hardware debuggers others emulate opentitan or related chips by their nature some of those drivers will maintain state in the rust struct that implements the transport trait others may have usb connections to hardware which gets disturbed if the usb connection is closed and restablished rust programs which link directly with opentitan lib are able to keep the transport object for a long as needed making a sequence of method invocations to stimulate the device under test however users of the command line opentitan tool does not have that ability since each command line invocation generally results in the creation of a transport object and the one useful method invocation before the transport is closed as the process terminates proposal we will add a new optional persistent “opentitan session proxy” process a daemon when used all communication with the transport backend will go through this proxy process the proxy process will listen on a tcp socket and contain the real transport implementation maintaining emulator state or connection to usb hardware the command line opentitan tool will use a stub transport implementation which forwards all requests via a tcp connection to the proxy design document tasks make sure all error return values from transport methods can be serialized with serde declare wire protocol by as a number of rust structs enums implement stub transport using serde json text via tcp implement proxy server functionality opentitansession binary ability for opentitansession to run in background as daemon this proposal has significant overlap with previous issue which proposed a json tcp protocol for communicating with an emulator running in a separate process with this proposal the interface to the emulator would instead be the transport trait itself that is the top level process of the emulator would either be opentitansession or would be another binary that links with a library containing most of the functionality of opentitansession the listening on tcp socket and translation of protocol into calls to transport trait | 0 |
17,168 | 11,756,107,318 | IssuesEvent | 2020-03-13 10:51:57 | CIRCL/AIL-framework | https://api.github.com/repos/CIRCL/AIL-framework | opened | Add an option in the onion explorer to hide the result when there is no screenshot | enhancement usability | Add an option in the onion explorer to hide the result when there is no screenshot | True | Add an option in the onion explorer to hide the result when there is no screenshot - Add an option in the onion explorer to hide the result when there is no screenshot | non_process | add an option in the onion explorer to hide the result when there is no screenshot add an option in the onion explorer to hide the result when there is no screenshot | 0 |
22,648 | 31,895,827,328 | IssuesEvent | 2023-09-18 01:31:56 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | Change term - formation | Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete | ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_formation
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation
* Term label (English, not normative): Formation
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| 1.0 | Change term - formation - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_formation
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation
* Term label (English, not normative): Formation
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| process | change term formation term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes formation term label english not normative formation organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the lithostratigraphic formation from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative notch peak formation house limestone fillmore formation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd | 1 |
10,720 | 13,523,058,801 | IssuesEvent | 2020-09-15 09:25:12 | pystatgen/sgkit | https://api.github.com/repos/pystatgen/sgkit | opened | Move IO repos to sgkit | multi-repo process + tools | We've [decided](https://discourse.pystatgen.org/t/20200910-developer-call-notes/96) to move sgkit-{bgen,plink,vcf} to the main sgkit repo. This is an umbrella issue to track tasks. See also discussion in #65
- [ ] Add Windows CI #247
- [ ] Move sgkit-bgen to main sgkit repo #256
- [ ] Move sgkit-plink to main sgkit repo #257
- [ ] Move sgkit-vcf to main sgkit repo #258
| 1.0 | Move IO repos to sgkit - We've [decided](https://discourse.pystatgen.org/t/20200910-developer-call-notes/96) to move sgkit-{bgen,plink,vcf} to the main sgkit repo. This is an umbrella issue to track tasks. See also discussion in #65
- [ ] Add Windows CI #247
- [ ] Move sgkit-bgen to main sgkit repo #256
- [ ] Move sgkit-plink to main sgkit repo #257
- [ ] Move sgkit-vcf to main sgkit repo #258
| process | move io repos to sgkit we ve to move sgkit bgen plink vcf to the main sgkit repo this is an umbrella issue to track tasks see also discussion in add windows ci move sgkit bgen to main sgkit repo move sgkit plink to main sgkit repo move sgkit vcf to main sgkit repo | 1 |
22,486 | 31,395,279,836 | IssuesEvent | 2023-08-26 21:22:57 | lynnandtonic/nestflix.fun | https://api.github.com/repos/lynnandtonic/nestflix.fun | closed | Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added) | suggested title in process | Please add as much of the following info as you can:
Title: Joan is Awful
Type (film/tv show): TV show - drama
Film or show in which it appears: Black Mirror
Is the parent film/show streaming anywhere? Yes - Netflix
About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful"
Actual footage of the film/show can be seen (yes/no)? Yes
Timestamps (interspersed with people watching it):
- 12:50 - 13:39
- 14:35 - 14:47
- 15:00 - 19:22
- 20:00 - 20:50
- 21:59 - 22:17
- 25:20 - 25:40
- 30:00 - 30:40
- 35:50 - 36:20
Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett
Network: Streamberry (Streamberry CEO: Mona Javadi)
Number of seasons: 1
Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing.

























 | 1.0 | Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added) - Please add as much of the following info as you can:
Title: Joan is Awful
Type (film/tv show): TV show - drama
Film or show in which it appears: Black Mirror
Is the parent film/show streaming anywhere? Yes - Netflix
About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful"
Actual footage of the film/show can be seen (yes/no)? Yes
Timestamps (interspersed with people watching it):
- 12:50 - 13:39
- 14:35 - 14:47
- 15:00 - 19:22
- 20:00 - 20:50
- 21:59 - 22:17
- 25:20 - 25:40
- 30:00 - 30:40
- 35:50 - 36:20
Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett
Network: Streamberry (Streamberry CEO: Mona Javadi)
Number of seasons: 1
Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing.

























 | process | add joan is awful from black mirror screenshots thumbnail and title card added please add as much of the following info as you can title joan is awful type film tv show tv show drama film or show in which it appears black mirror is the parent film show streaming anywhere yes netflix about when in the parent film show does it appear episode joan is awful actual footage of the film show can be seen yes no yes timestamps interspersed with people watching it starring salma hayek himesh patel jaboukie young white camirin farmer ben marnes cate blanchett network streamberry streamberry ceo mona javadi number of seasons note this episode of black mirror is crazy meta there s a version of joan is awful within the show of joan is awful and the entirity of the episode is revealed to be the real joan is awful show i m just putting the primary joan is awful show here in order to not be too confusing | 1 |
12,692 | 15,058,119,550 | IssuesEvent | 2021-02-03 22:51:03 | panther-labs/panther | https://api.github.com/repos/panther-labs/panther | opened | Standard Fields - Edge Cases and Oddities | bug team:data processing | ### Describe the bug
Reviewing the schemas revealed some oddities and edge cases for the Standard Fields that Panther normalizes across all log records.
### Expected behavior
The following are summaries of the anomalous behavior from the `log_processor` testdata.
Refer to the **Additional Context** section for more details.
#### Anomalous Behavior
- GitLab - Production Tests<br />The `location` field (`"http://34.222.254.254/users/sign_in"`) shows an IP Address. <br />Shouldn't `p_any_ip_addresses` include `34.222.254.254`?<br /><br />
- GSuite - Data clearly shows domain names: <br />i.e. `www.google.com/accounts/OAuthLogin` and `77185425430.apps.googleusercontent.com` <br />However the results do not include any `p_any_domain_names`<br /><br />
- AWS - These values should be included in their respective `p_` fields but are not included:
- `p_any_aws_arns` - `arn:aws:wafv2:us-west-2:123456789EXAMPLE:regional/webacl/panther_web_dev/f96fe2f4-a2f4-4c06-a4bc-afd495bd675f`
- `p_any_domain_names` - `web-123456789.us-west-2.elb.amazonaws.com`
- `p_any_aws_account_ids` - `123456789EXAMPLE` and `123456789EXAMPLE_MANAGED`
#### Others
- Consider `ip-172-31-14-137.us-west-2.compute.internal` <br />While this is explicitly a domain name, it also explicitly denotes an IP address. <br />Would we want to extract `172.31.14.137` as an IP address?<br /><br />
- From the Sophos logs:
- There is a file path value: `"/Users/jsmith/Dropbox/.dropbox.cache/new_files/eec8381bba1914aa2207f4a53c0fad1e"`. <br />Based on the context of the test data, this appears to be a Win x32 system. <br />However this file path is not a valid Windows file path (would be valid if `/` were `\`)
- In one of the tests, there is a file path explicitly stated in the `name`. <br />Since the file path input key is not present, the results do not show any file paths. <br />Would this make a case to add support for file paths in the standardized `p_` fields?
- AWS
- VPC DNS Tests
- `p_any_domain_names`
- Why is `xn--fa-hia.com` not included in this `p_` field?
- What's up with the `"faß.com"` included in this `p_` field?
- Why is `foo@bar.com` included in this `p_` field?
- Shouldn't `foo@bar.com` be included in the `p_any_emails` field?
- CloudTrail (base and insight) Tests
- While not particularly useful, shouldn't `p_any_domain_names` match on the following:
- `kms.amazonaws.com`
- `ssm.amazonaws.com`
### Additional context
See this [Pull Request Review](https://github.com/panther-labs/panther-analysis/pull/187#pullrequestreview-581874346) for full context. | 1.0 | Standard Fields - Edge Cases and Oddities - ### Describe the bug
Reviewing the schemas revealed some oddities and edge cases for the Standard Fields that Panther normalizes across all log records.
### Expected behavior
The following are summaries of the anomalous behavior from the `log_processor` testdata.
Refer to the **Additional Context** section for more details.
#### Anomalous Behavior
- GitLab - Production Tests<br />The `location` field (`"http://34.222.254.254/users/sign_in"`) shows an IP Address. <br />Shouldn't `p_any_ip_addresses` include `34.222.254.254`?<br /><br />
- GSuite - Data clearly shows domain names: <br />i.e. `www.google.com/accounts/OAuthLogin` and `77185425430.apps.googleusercontent.com` <br />However the results do not include any `p_any_domain_names`<br /><br />
- AWS - These values should be included in their respective `p_` fields but are not included:
- `p_any_aws_arns` - `arn:aws:wafv2:us-west-2:123456789EXAMPLE:regional/webacl/panther_web_dev/f96fe2f4-a2f4-4c06-a4bc-afd495bd675f`
- `p_any_domain_names` - `web-123456789.us-west-2.elb.amazonaws.com`
- `p_any_aws_account_ids` - `123456789EXAMPLE` and `123456789EXAMPLE_MANAGED`
#### Others
- Consider `ip-172-31-14-137.us-west-2.compute.internal` <br />While this is explicitly a domain name, it also explicitly denotes an IP address. <br />Would we want to extract `172.31.14.137` as an IP address?<br /><br />
- From the Sophos logs:
- There is a file path value: `"/Users/jsmith/Dropbox/.dropbox.cache/new_files/eec8381bba1914aa2207f4a53c0fad1e"`. <br />Based on the context of the test data, this appears to be a Win x32 system. <br />However this file path is not a valid Windows file path (would be valid if `/` were `\`)
- In one of the tests, there is a file path explicitly stated in the `name`. <br />Since the file path input key is not present, the results do not show any file paths. <br />Would this make a case to add support for file paths in the standardized `p_` fields?
- AWS
- VPC DNS Tests
- `p_any_domain_names`
- Why is `xn--fa-hia.com` not included in this `p_` field?
- What's up with the `"faß.com"` included in this `p_` field?
- Why is `foo@bar.com` included in this `p_` field?
- Shouldn't `foo@bar.com` be included in the `p_any_emails` field?
- CloudTrail (base and insight) Tests
- While not particularly useful, shouldn't `p_any_domain_names` match on the following:
- `kms.amazonaws.com`
- `ssm.amazonaws.com`
### Additional context
See this [Pull Request Review](https://github.com/panther-labs/panther-analysis/pull/187#pullrequestreview-581874346) for full context. | process | standard fields edge cases and oddities describe the bug reviewing the schemas revealed some oddities and edge cases for the standard fields that panther normalizes across all log records expected behavior the following are summaries of the anomalous behavior from the log processor testdata refer to the additional context section for more details anomalous behavior gitlab production tests the location field shows an ip address shouldn t p any ip addresses include gsuite data clearly shows domain names i e and apps googleusercontent com however the results do not include any p any domain names aws these values should be included in their respective p fields but are not included p any aws arns arn aws us west regional webacl panther web dev p any domain names web us west elb amazonaws com p any aws account ids and managed others consider ip us west compute internal while this is explicitly a domain name it also explicitly denotes an ip address would we want to extract as an ip address from the sophos logs there is a file path value users jsmith dropbox dropbox cache new files based on the context of the test data this appears to be a win system however this file path is not a valid windows file path would be valid if were in one of the tests there is a file path explicitly stated in the name since the file path input key is not present the results do not show any file paths would this make a case to add support for file paths in the standardized p fields aws vpc dns tests p any domain names why is xn fa hia com not included in this p field what s up with the faß com included in this p field why is foo bar com included in this p field shouldn t foo bar com be included in the p any emails field cloudtrail base and insight tests while not particularly useful shouldn t p any domain names match on the following kms amazonaws com ssm amazonaws com additional context see this for full context | 1 |
2,836 | 5,791,834,513 | IssuesEvent | 2017-05-02 07:36:55 | gaocegege/maintainer | https://api.github.com/repos/gaocegege/maintainer | opened | Transfer from builder pattern to multi-stage docker build | priority/P3 process/not claimed type/enhancement | ref http://blog.alexellis.io/mutli-stage-docker-builds/
Blocked until the feature is released.
ref https://github.com/moby/moby/pull/31257 and https://github.com/moby/moby/pull/32063
| 1.0 | Transfer from builder pattern to multi-stage docker build - ref http://blog.alexellis.io/mutli-stage-docker-builds/
Blocked until the feature is released.
ref https://github.com/moby/moby/pull/31257 and https://github.com/moby/moby/pull/32063
| process | transfer from builder pattern to multi stage docker build ref blocked until the feature is released ref and | 1 |
120,708 | 25,850,910,844 | IssuesEvent | 2022-12-13 10:20:03 | Clueless-Community/seamless-ui | https://api.github.com/repos/Clueless-Community/seamless-ui | closed | Improve Footer Dark 4 | MEDIUM codepeak 22 issue:3 | Need to improve this component on `Footer/src/footer-dark-04.html`
## Improvements required :
The Button, Spacing, social media icons are not aligned as per figma file
## Any reference image?

| 1.0 | Improve Footer Dark 4 - Need to improve this component on `Footer/src/footer-dark-04.html`
## Improvements required :
The Button, Spacing, social media icons are not aligned as per figma file
## Any reference image?

| non_process | improve footer dark need to improve this component on footer src footer dark html improvements required the button spacing social media icons are not aligned as per figma file any reference image | 0 |
21,915 | 30,443,576,963 | IssuesEvent | 2023-07-15 11:29:14 | h4sh5/pypi-auto-scanner | https://api.github.com/repos/h4sh5/pypi-auto-scanner | opened | skypilot-nightly 1.0.0.dev20230715 has 2 GuardDog issues | guarddog exec-base64 silent-process-execution | https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230715",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230715/sky/cloud_stores.py:110",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230715/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp79q_y85t/skypilot-nightly"
}
}``` | 1.0 | skypilot-nightly 1.0.0.dev20230715 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230715",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230715/sky/cloud_stores.py:110",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230715/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp79q_y85t/skypilot-nightly"
}
}``` | process | skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp skypilot nightly | 1 |
760,363 | 26,638,008,781 | IssuesEvent | 2023-01-25 00:17:22 | ArjunSharda/Passeo | https://api.github.com/repos/ArjunSharda/Passeo | closed | Add more security to Passeo quickgenerate command | enhancement High Priority | This update will add more security to Passeo's quickgenerate's command. | 1.0 | Add more security to Passeo quickgenerate command - This update will add more security to Passeo's quickgenerate's command. | non_process | add more security to passeo quickgenerate command this update will add more security to passeo s quickgenerate s command | 0 |
53,500 | 3,040,705,628 | IssuesEvent | 2015-08-07 16:52:11 | scamille/simc_issue_test3 | https://api.github.com/repos/scamille/simc_issue_test3 | closed | Non-zero melee stat scale factors for Warlocks | bug imported Priority-Medium | _From [natehieter@gmail.com](https://code.google.com/u/natehieter@gmail.com/) on January 09, 2009 22:21:39_
When many Warlocks are added to the sim, some end up with apparent scaling
against stats like expertise. Most likely, this is due to a spike in one
or more iterations due to a bug.
_Original issue: http://code.google.com/p/simulationcraft/issues/detail?id=18_ | 1.0 | Non-zero melee stat scale factors for Warlocks - _From [natehieter@gmail.com](https://code.google.com/u/natehieter@gmail.com/) on January 09, 2009 22:21:39_
When many Warlocks are added to the sim, some end up with apparent scaling
against stats like expertise. Most likely, this is due to a spike in one
or more iterations due to a bug.
_Original issue: http://code.google.com/p/simulationcraft/issues/detail?id=18_ | non_process | non zero melee stat scale factors for warlocks from on january when many warlocks are added to the sim some end up with apparent scaling against stats like expertise most likely this is due to a spike in one or more iterations due to a bug original issue | 0 |
9,616 | 12,553,266,190 | IssuesEvent | 2020-06-06 21:16:53 | peopledoc/procrastinate | https://api.github.com/repos/peopledoc/procrastinate | closed | Auto-deploy to PyPI is still broken | Good for: newcomers Type: Bug Type: Process | It looks like our .travis.yml syntax is still broken, due to a mix between matrix and jobs. | 1.0 | Auto-deploy to PyPI is still broken - It looks like our .travis.yml syntax is still broken, due to a mix between matrix and jobs. | process | auto deploy to pypi is still broken it looks like our travis yml syntax is still broken due to a mix between matrix and jobs | 1 |
1,565 | 10,343,139,407 | IssuesEvent | 2019-09-04 08:18:25 | DimensionDev/Maskbook | https://api.github.com/repos/DimensionDev/Maskbook | closed | Just script: use of 'shell: true' | Dev: Local Automation Severity: Low Speedy Close Requested Type: Bug wontfix | > without this things won't work but it said this option is dangerous.
-- @Artoria2e5 | 1.0 | Just script: use of 'shell: true' - > without this things won't work but it said this option is dangerous.
-- @Artoria2e5 | non_process | just script use of shell true without this things won t work but it said this option is dangerous | 0 |
147,264 | 11,780,316,280 | IssuesEvent | 2020-03-16 19:48:24 | theophilusx/ssh2-sftp-client | https://api.github.com/repos/theophilusx/ssh2-sftp-client | closed | Reconnect on connection closed | bug test | We are building an upload tool that must be resilient to connection loss. Right now when the connection is lost while an append or put operation is in progress, we get a "close" event, but the append or put operations don't resolve or throw. Currently I'm wrapping our upload as such:
```
try {
const success = await new Promise<boolean>(resolv => {
const listener = (s: boolean) => {
this.client.removeListener("close", listener);
resolv(s);
};
this.client.on("close", () => listener(false));
// If the connection is lost, neither of these ever resolve.
if (tryCont) {
this.client.append(stream, destFile).then(() => listener(true));
} else {
this.client.put(stream, destFile).then(() => listener(true));
}
});
if (!success) throw new Error("Connection lost");
} catch (e) {
console.log(e);
throw e;
}
```
This works, but feels like its leaking in the append and put calls since they are not resolving.
Along the same lines, calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of "no more sessons" errors when a connection is finally made. Its like each failed connect call finally makes a connection even when it reports failure when its initially called. Is there an intended approach for doing this kind of thing? Thanks! | 1.0 | Reconnect on connection closed - We are building an upload tool that must be resilient to connection loss. Right now when the connection is lost while an append or put operation is in progress, we get a "close" event, but the append or put operations don't resolve or throw. Currently I'm wrapping our upload as such:
```
try {
const success = await new Promise<boolean>(resolv => {
const listener = (s: boolean) => {
this.client.removeListener("close", listener);
resolv(s);
};
this.client.on("close", () => listener(false));
// If the connection is lost, neither of these ever resolve.
if (tryCont) {
this.client.append(stream, destFile).then(() => listener(true));
} else {
this.client.put(stream, destFile).then(() => listener(true));
}
});
if (!success) throw new Error("Connection lost");
} catch (e) {
console.log(e);
throw e;
}
```
This works, but feels like its leaking in the append and put calls since they are not resolving.
Along the same lines, calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of "no more sessons" errors when a connection is finally made. Its like each failed connect call finally makes a connection even when it reports failure when its initially called. Is there an intended approach for doing this kind of thing? Thanks! | non_process | reconnect on connection closed we are building an upload tool that must be resilient to connection loss right now when the connection is lost while an append or put operation is in progress we get a close event but the append or put operations don t resolve or throw currently i m wrapping our upload as such try const success await new promise resolv const listener s boolean this client removelistener close listener resolv s this client on close listener false if the connection is lost neither of these ever resolve if trycont this client append stream destfile then listener true else this client put stream destfile then listener true if success throw new error connection lost catch e console log e throw e this works but feels like its leaking in the append and put calls since they are not resolving along the same lines calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of no more sessons errors when a connection is finally made its like each failed connect call finally makes a connection even when it reports failure when its initially called is there an intended approach for doing this kind of thing thanks | 0 |
20,188 | 26,753,344,520 | IssuesEvent | 2023-01-30 21:33:43 | darkside-princeton/sipm-analysis | https://api.github.com/repos/darkside-princeton/sipm-analysis | closed | Rename SiPM and Dataset classes | pre-processing | Rename class SiPM to WaveformAnalyzer and class Dataset to WaveformDataset. Modify all files accordingly. | 1.0 | Rename SiPM and Dataset classes - Rename class SiPM to WaveformAnalyzer and class Dataset to WaveformDataset. Modify all files accordingly. | process | rename sipm and dataset classes rename class sipm to waveformanalyzer and class dataset to waveformdataset modify all files accordingly | 1 |
35,527 | 4,995,422,933 | IssuesEvent | 2016-12-09 10:05:51 | halestudio/hale | https://api.github.com/repos/halestudio/hale | closed | Arc interpolation takes very long or is not terminating | bug io to be tested | I have issues with the arc interpolation taking seemingly forever. I was not able to complete loading a project because loading the data did not finish in a sensible amount of time.
refs #181
Internal example project `aaa-au-flurstuecke`. Please create a test case from the respective geometry causing this problem. | 1.0 | Arc interpolation takes very long or is not terminating - I have issues with the arc interpolation taking seemingly forever. I was not able to complete loading a project because loading the data did not finish in a sensible amount of time.
refs #181
Internal example project `aaa-au-flurstuecke`. Please create a test case from the respective geometry causing this problem. | non_process | arc interpolation takes very long or is not terminating i have issues with the arc interpolation taking seemingly forever i was not able to complete loading a project because loading the data did not finish in a sensible amount of time refs internal example project aaa au flurstuecke please create a test case from the respective geometry causing this problem | 0 |
162,379 | 20,188,185,937 | IssuesEvent | 2022-02-11 01:16:15 | Nivaskumark/kernel_v4.19.72_old | https://api.github.com/repos/Nivaskumark/kernel_v4.19.72_old | opened | CVE-2022-0492 (High) detected in linuxlinux-4.19.83 | security vulnerability | ## CVE-2022-0492 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The cgroup release_agent is called with call_usermodehelper. The function
call_usermodehelper starts the release_agent with a full set fo capabilities.
Therefore require capabilities when setting the release_agaent.
Before version v5.17-rc3
<p>Publish Date: 2022-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0492>CVE-2022-0492</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-0492">https://security-tracker.debian.org/tracker/CVE-2022-0492</a></p>
<p>Release Date: 2022-02-04</p>
<p>Fix Resolution: v5.17-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0492 (High) detected in linuxlinux-4.19.83 - ## CVE-2022-0492 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The cgroup release_agent is called with call_usermodehelper. The function
call_usermodehelper starts the release_agent with a full set fo capabilities.
Therefore require capabilities when setting the release_agaent.
Before version v5.17-rc3
<p>Publish Date: 2022-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0492>CVE-2022-0492</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-0492">https://security-tracker.debian.org/tracker/CVE-2022-0492</a></p>
<p>Release Date: 2022-02-04</p>
<p>Fix Resolution: v5.17-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files kernel cgroup cgroup c kernel cgroup cgroup c vulnerability details the cgroup release agent is called with call usermodehelper the function call usermodehelper starts the release agent with a full set fo capabilities therefore require capabilities when setting the release agaent before version publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
763,027 | 26,741,230,452 | IssuesEvent | 2023-01-30 13:08:15 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Provide custom host verifier implementation for pgjdbc | kind/bug priority/medium area/ecosystem | Jira Link: [DB-3807](https://yugabyte.atlassian.net/browse/DB-3807)
### Description
In a YB Managed setup where a client is in a VPC peered with the YBDB cluster's VPC, the ssl connection with mode `verify-full` does not work.
Provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by `yb_servers()` function. | 1.0 | [YSQL] Provide custom host verifier implementation for pgjdbc - Jira Link: [DB-3807](https://yugabyte.atlassian.net/browse/DB-3807)
### Description
In a YB Managed setup where a client is in a VPC peered with the YBDB cluster's VPC, the ssl connection with mode `verify-full` does not work.
Provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by `yb_servers()` function. | non_process | provide custom host verifier implementation for pgjdbc jira link description in a yb managed setup where a client is in a vpc peered with the ybdb cluster s vpc the ssl connection with mode verify full does not work provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by yb servers function | 0 |
236,013 | 25,971,399,283 | IssuesEvent | 2022-12-19 11:31:32 | nk7598/linux-4.19.72 | https://api.github.com/repos/nk7598/linux-4.19.72 | closed | CVE-2022-1998 (High) detected in linuxlinux-4.19.269 - autoclosed | security vulnerability | ## CVE-2022-1998 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nk7598/linux-4.19.72/commit/8d6de636016872da224f31e7d9d0fe96d373b46c">8d6de636016872da224f31e7d9d0fe96d373b46c</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system.
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-1998 (High) detected in linuxlinux-4.19.269 - autoclosed - ## CVE-2022-1998 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nk7598/linux-4.19.72/commit/8d6de636016872da224f31e7d9d0fe96d373b46c">8d6de636016872da224f31e7d9d0fe96d373b46c</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system.
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href vulnerable source files fs notify fanotify fanotify user c fs notify fanotify fanotify user c vulnerability details a use after free in the linux kernel file system notify functionality was found in the way user triggers copy info records to user call to fail in copy event to user a local user could use this flaw to crash the system or potentially escalate their privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
18,258 | 24,341,318,210 | IssuesEvent | 2022-10-01 18:47:49 | RobertCraigie/prisma-client-py | https://api.github.com/repos/RobertCraigie/prisma-client-py | closed | Argument list too long when connecting to a database with a large schema | bug/0-needs-info kind/bug process/candidate topic: client level/intermediate priority/high | ## Bug description
The query engine cannot spawn the query engine child process when the schema is too large.
The error is coming from that line https://github.com/RobertCraigie/prisma-client-py/blob/main/src/prisma/generator/templates/engine/query.py.jinja#L106. When the schema is too large, it's content will be truncated by the OS and the engine will fail to start. The following patch will circumvent the issue by passing the path to the schema instead of it's content:
```
env = os.environ.copy()
env.update(
# PRISMA_DML=self.dml,
RUST_LOG='error',
RUST_LOG_FORMAT='json',
PRISMA_CLIENT_ENGINE_TYPE='binary',
)
if os.environ.get('PRISMA_DML_PATH'):
env.update(PRISMA_DML_PATH=os.environ.get('PRISMA_DML_PATH'))
else:
env.update(PRISMA_DML=self.dml)
```
## How to reproduce
Steps to reproduce the behavior:
1. Generate a prisma.schema file with a large number of tables and fields
2. Generate the prisma client
3. Connect to the database
4. See error:
```
Traceback (most recent call last):
File "...", line 239, in <module>
app = CashflowLoader(args.process_date)
File "...", line 35, in __init__
self.db.connect()
File ".../venv/lib/python3.9/site-packages/prisma/client.py", line 3657, in connect
self.__engine.connect(
File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 110, in connect
self.spawn(file, timeout=timeout, datasources=datasources)
File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 152, in spawn
self.process = subprocess.Popen(
File ".../lib/python3.9/subprocess.py", line 951, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File ".../lib/python3.9/subprocess.py", line 1821, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
OSError: [Errno 7] Argument list too long: '/tmp/prisma/binaries/engines/efdf9b1183dddfd4258cd181a72125755215ab7b/prisma-query-engine-debian-openssl-1.1.x'
```
## Expected behavior
No error should be reported and the connexion should be established with the database.
## Prisma information
Create a large schema file
No specific queries required
## Environment & setup
- OS: Windows 11, Ubuntu 20.10
- Database: MariaDB
- Python version: 3.9.7
- Prisma version:
```
prisma : 3.13.0
prisma client python : 0.6.6
platform : debian-openssl-1.1.x
engines : efdf9b1183dddfd4258cd181a72125755215ab7b
install path : /mnt/c/Users/andrew/OneDrive/Git_Repos/e22/Benzaiten_repos/benzaiten-processes/venv/lib/python3.9/site-packages/prisma
installed extras : []
```
| 1.0 | Argument list too long when connecting to a database with a large schema - ## Bug description
The query engine cannot spawn the query engine child process when the schema is too large.
The error is coming from that line https://github.com/RobertCraigie/prisma-client-py/blob/main/src/prisma/generator/templates/engine/query.py.jinja#L106. When the schema is too large, it's content will be truncated by the OS and the engine will fail to start. The following patch will circumvent the issue by passing the path to the schema instead of it's content:
```
env = os.environ.copy()
env.update(
# PRISMA_DML=self.dml,
RUST_LOG='error',
RUST_LOG_FORMAT='json',
PRISMA_CLIENT_ENGINE_TYPE='binary',
)
if os.environ.get('PRISMA_DML_PATH'):
env.update(PRISMA_DML_PATH=os.environ.get('PRISMA_DML_PATH'))
else:
env.update(PRISMA_DML=self.dml)
```
## How to reproduce
Steps to reproduce the behavior:
1. Generate a prisma.schema file with a large number of tables and fields
2. Generate the prisma client
3. Connect to the database
4. See error:
```
Traceback (most recent call last):
File "...", line 239, in <module>
app = CashflowLoader(args.process_date)
File "...", line 35, in __init__
self.db.connect()
File ".../venv/lib/python3.9/site-packages/prisma/client.py", line 3657, in connect
self.__engine.connect(
File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 110, in connect
self.spawn(file, timeout=timeout, datasources=datasources)
File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 152, in spawn
self.process = subprocess.Popen(
File ".../lib/python3.9/subprocess.py", line 951, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File ".../lib/python3.9/subprocess.py", line 1821, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
OSError: [Errno 7] Argument list too long: '/tmp/prisma/binaries/engines/efdf9b1183dddfd4258cd181a72125755215ab7b/prisma-query-engine-debian-openssl-1.1.x'
```
## Expected behavior
No error should be reported and the connexion should be established with the database.
## Prisma information
Create a large schema file
No specific queries required
## Environment & setup
- OS: Windows 11, Ubuntu 20.10
- Database: MariaDB
- Python version: 3.9.7
- Prisma version:
```
prisma : 3.13.0
prisma client python : 0.6.6
platform : debian-openssl-1.1.x
engines : efdf9b1183dddfd4258cd181a72125755215ab7b
install path : /mnt/c/Users/andrew/OneDrive/Git_Repos/e22/Benzaiten_repos/benzaiten-processes/venv/lib/python3.9/site-packages/prisma
installed extras : []
```
| process | argument list too long when connecting to a database with a large schema bug description the query engine cannot spawn the query engine child process when the schema is too large the error is coming from that line when the schema is too large it s content will be truncated by the os and the engine will fail to start the following patch will circumvent the issue by passing the path to the schema instead of it s content env os environ copy env update prisma dml self dml rust log error rust log format json prisma client engine type binary if os environ get prisma dml path env update prisma dml path os environ get prisma dml path else env update prisma dml self dml how to reproduce steps to reproduce the behavior generate a prisma schema file with a large number of tables and fields generate the prisma client connect to the database see error traceback most recent call last file line in app cashflowloader args process date file line in init self db connect file venv lib site packages prisma client py line in connect self engine connect file venv lib site packages prisma engine query py line in connect self spawn file timeout timeout datasources datasources file venv lib site packages prisma engine query py line in spawn self process subprocess popen file lib subprocess py line in init self execute child args executable preexec fn close fds file lib subprocess py line in execute child raise child exception type errno num err msg err filename oserror argument list too long tmp prisma binaries engines prisma query engine debian openssl x expected behavior no error should be reported and the connexion should be established with the database prisma information create a large schema file no specific queries required environment setup os windows ubuntu database mariadb python version prisma version prisma prisma client python platform debian openssl x engines install path mnt c users andrew onedrive git repos benzaiten repos benzaiten processes venv lib site packages prisma installed extras | 1 |
15,229 | 19,100,935,004 | IssuesEvent | 2021-11-29 22:26:47 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | Bazel wants full dependency graph for incompatible targets - why aren't they just skipped? | type: support / not a bug (process) team-Configurability untriaged | ### Description of the problem / feature request:
Hi!
I have a target that is only compatible with the `aarch64` platform:
```
cc_binary(
name = "foo",
target_compatible_with = ["@platforms//cpu:aarch64"],
deps = [ "//foo:bar", ]
)
```
Now, the `bar` dependency comes in two flavors, so I have an alias:
```
alias(
name = "bar",
actual = select({
# Simplified syntax here
"aarch64_linux" : ":bar_linux",
"aarch64_qnx": ":bar_qnx",
})
)
```
Now, my problem is that when I build for `x86` using wildcard, Bazel complains that it cannot find any `bar` for x86. And that's of course expected - there's no `bar` or x86! However why is Bazel complaining here? I specified that this target is incompatible with `x86`, so it shouldn't need to analyze it. Why is it doing that?
If I remove the `bar` dependency, Bazel just skips the target and exits gracefully without errors.
I could solve this problem adding a default condition for `bar`, but I don't want that. It will silence a useful warning. If I ever want to support a 3rd `aarch64` version, the default condition will silence the warning without prompting me to add a 3rd `aarch64` version of `bar`.
I also don't want to add a `select` in the `deps` field of the `cc_binary` either - it's redundant and adds noise. I already specify that the target is only compatible with`aarch64`. I don't need to add the `deps` only for `aarch64`.
Is this expected? Is there any other way I could solve this problem?
Thanks!
### Feature requests: what underlying problem are you trying to solve with this feature?
I want Bazel to ignore incompatible targets, without it trying to find its dependencies (which may not exist at all for an incompatible platform)
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
See above.
### What operating system are you running Bazel on?
Ubuntu 18.04
### What's the output of `bazel info release`?
release 5.0.0-pre.20210907.1
### Have you found anything relevant by searching the web?
Nothing on the web. Asked in [bazel-discuss](https://groups.google.com/g/bazel-discuss/c/NNlN2Jl_bWI) without answer.
| 1.0 | Bazel wants full dependency graph for incompatible targets - why aren't they just skipped? - ### Description of the problem / feature request:
Hi!
I have a target that is only compatible with the `aarch64` platform:
```
cc_binary(
name = "foo",
target_compatible_with = ["@platforms//cpu:aarch64"],
deps = [ "//foo:bar", ]
)
```
Now, the `bar` dependency comes in two flavors, so I have an alias:
```
alias(
name = "bar",
actual = select({
# Simplified syntax here
"aarch64_linux" : ":bar_linux",
"aarch64_qnx": ":bar_qnx",
})
)
```
Now, my problem is that when I build for `x86` using wildcard, Bazel complains that it cannot find any `bar` for x86. And that's of course expected - there's no `bar` or x86! However why is Bazel complaining here? I specified that this target is incompatible with `x86`, so it shouldn't need to analyze it. Why is it doing that?
If I remove the `bar` dependency, Bazel just skips the target and exits gracefully without errors.
I could solve this problem adding a default condition for `bar`, but I don't want that. It will silence a useful warning. If I ever want to support a 3rd `aarch64` version, the default condition will silence the warning without prompting me to add a 3rd `aarch64` version of `bar`.
I also don't want to add a `select` in the `deps` field of the `cc_binary` either - it's redundant and adds noise. I already specify that the target is only compatible with`aarch64`. I don't need to add the `deps` only for `aarch64`.
Is this expected? Is there any other way I could solve this problem?
Thanks!
### Feature requests: what underlying problem are you trying to solve with this feature?
I want Bazel to ignore incompatible targets, without it trying to find its dependencies (which may not exist at all for an incompatible platform)
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
See above.
### What operating system are you running Bazel on?
Ubuntu 18.04
### What's the output of `bazel info release`?
release 5.0.0-pre.20210907.1
### Have you found anything relevant by searching the web?
Nothing on the web. Asked in [bazel-discuss](https://groups.google.com/g/bazel-discuss/c/NNlN2Jl_bWI) without answer.
| process | bazel wants full dependency graph for incompatible targets why aren t they just skipped description of the problem feature request hi i have a target that is only compatible with the platform cc binary name foo target compatible with deps now the bar dependency comes in two flavors so i have an alias alias name bar actual select simplified syntax here linux bar linux qnx bar qnx now my problem is that when i build for using wildcard bazel complains that it cannot find any bar for and that s of course expected there s no bar or however why is bazel complaining here i specified that this target is incompatible with so it shouldn t need to analyze it why is it doing that if i remove the bar dependency bazel just skips the target and exits gracefully without errors i could solve this problem adding a default condition for bar but i don t want that it will silence a useful warning if i ever want to support a version the default condition will silence the warning without prompting me to add a version of bar i also don t want to add a select in the deps field of the cc binary either it s redundant and adds noise i already specify that the target is only compatible with i don t need to add the deps only for is this expected is there any other way i could solve this problem thanks feature requests what underlying problem are you trying to solve with this feature i want bazel to ignore incompatible targets without it trying to find its dependencies which may not exist at all for an incompatible platform bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible see above what operating system are you running bazel on ubuntu what s the output of bazel info release release pre have you found anything relevant by searching the web nothing on the web asked in without answer | 1 |
415,403 | 12,128,983,184 | IssuesEvent | 2020-04-22 21:33:16 | microsoftgraph/microsoft-graph-toolkit | https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit | closed | [BUG] In person initials non English letters get ignored | Priority: 0 State: Completed State: Started bug work-in-progress | <!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. -->
<!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. -->
**Describe the bug**
We filter out non A-Z characters in the initials, should we not use more intelligent regex to filter out bad characters (assumedly numbers, and punctuation etc.). A quick google suggests \p{L} might work.
**To Reproduce**
Steps to reproduce the behavior:
1. Use the person card with a user whose first or second name starts with initials that aren't a-z. E.g Ægir Davidsson and doesn't have an avatar
2. View initials image
3. See initials as just D
4. JSFiddle: https://jsfiddle.net/robpethi/a08dwyh7/3/
<!-- if possible, please link to a GitHub project that reproduces the issue -->
**Expected behavior**
Expected initials to be ÆD
**Environment (please complete the following information):**
- Framework react
- Context Teams
- Version Latest
| 1.0 | [BUG] In person initials non English letters get ignored - <!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. -->
<!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. -->
**Describe the bug**
We filter out non A-Z characters in the initials, should we not use more intelligent regex to filter out bad characters (assumedly numbers, and punctuation etc.). A quick google suggests \p{L} might work.
**To Reproduce**
Steps to reproduce the behavior:
1. Use the person card with a user whose first or second name starts with initials that aren't a-z. E.g Ægir Davidsson and doesn't have an avatar
2. View initials image
3. See initials as just D
4. JSFiddle: https://jsfiddle.net/robpethi/a08dwyh7/3/
<!-- if possible, please link to a GitHub project that reproduces the issue -->
**Expected behavior**
Expected initials to be ÆD
**Environment (please complete the following information):**
- Framework react
- Context Teams
- Version Latest
| non_process | in person initials non english letters get ignored describe the bug we filter out non a z characters in the initials should we not use more intelligent regex to filter out bad characters assumedly numbers and punctuation etc a quick google suggests p l might work to reproduce steps to reproduce the behavior use the person card with a user whose first or second name starts with initials that aren t a z e g ægir davidsson and doesn t have an avatar view initials image see initials as just d jsfiddle expected behavior expected initials to be æd environment please complete the following information framework react context teams version latest | 0 |
192,576 | 22,215,973,194 | IssuesEvent | 2022-06-08 01:42:43 | artsking/linux-4.1.15 | https://api.github.com/repos/artsking/linux-4.1.15 | reopened | CVE-2016-4470 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2016-4470 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command.
<p>Publish Date: 2016-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4470>CVE-2016-4470</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p>
<p>Release Date: 2016-06-27</p>
<p>Fix Resolution: v4.7-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-4470 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-4470 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command.
<p>Publish Date: 2016-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4470>CVE-2016-4470</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p>
<p>Release Date: 2016-06-27</p>
<p>Fix Resolution: v4.7-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the key reject and link function in security keys key c in the linux kernel through does not ensure that a certain data structure is initialized which allows local users to cause a denial of service system crash via vectors involving a crafted keyctl command publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
19,691 | 26,045,591,858 | IssuesEvent | 2022-12-22 14:08:53 | aiidateam/aiida-core | https://api.github.com/repos/aiidateam/aiida-core | closed | Allow user to display meaning of exit status in `verdi process list` | topic/verdi type/accepted feature topic/processes | ### Is your feature request related to a problem? Please describe
`verdi process list` presents information about why calculations fail:
```
$ verdi process list -a -L PwCalculation
PK Created Process label Process State Process status
---- --------- --------------- ---------------- ---------------------
194 21h ago PwCalculation ☠ Killed Killed by parent<186>
243 20h ago PwCalculation ⏹ Finished [410]
280 20h ago PwCalculation ☠ Killed Killed by parent<263>
287 20h ago PwCalculation ☠ Killed Killed by parent<282>
733 18h ago PwCalculation ⏹ Finished [0]
741 18h ago PwCalculation ⏹ Finished [400]
785 16h ago PwCalculation ⏹ Finished [0]
790 16h ago PwCalculation ⏹ Finished [312]
834 16h ago PwCalculation ⏹ Finished [340]
837 16h ago PwCalculation ⏹ Finished [0]
856 13h ago PwCalculation ⏹ Finished [305]
873 13h ago PwCalculation ⏹ Finished [0]
887 12h ago PwCalculation ⏹ Finished [305]
903 12h ago PwCalculation ⏹ Finished [305]
```
Some of it is self-explanatory ("Killed"), but the exit codes are not (and they depend on the plugin, so it's not enough to "learn them once").
In order to figure out what they mean, a new user has to go through the reports of the individual calculations.
### Describe the solution you'd like
It would be great if a user could request to print the message associated with the exit status as a last column in `verdi process list`.
@sphuber What do you think?
| 1.0 | Allow user to display meaning of exit status in `verdi process list` - ### Is your feature request related to a problem? Please describe
`verdi process list` presents information about why calculations fail:
```
$ verdi process list -a -L PwCalculation
PK Created Process label Process State Process status
---- --------- --------------- ---------------- ---------------------
194 21h ago PwCalculation ☠ Killed Killed by parent<186>
243 20h ago PwCalculation ⏹ Finished [410]
280 20h ago PwCalculation ☠ Killed Killed by parent<263>
287 20h ago PwCalculation ☠ Killed Killed by parent<282>
733 18h ago PwCalculation ⏹ Finished [0]
741 18h ago PwCalculation ⏹ Finished [400]
785 16h ago PwCalculation ⏹ Finished [0]
790 16h ago PwCalculation ⏹ Finished [312]
834 16h ago PwCalculation ⏹ Finished [340]
837 16h ago PwCalculation ⏹ Finished [0]
856 13h ago PwCalculation ⏹ Finished [305]
873 13h ago PwCalculation ⏹ Finished [0]
887 12h ago PwCalculation ⏹ Finished [305]
903 12h ago PwCalculation ⏹ Finished [305]
```
Some of it is self-explanatory ("Killed"), but the exit codes are not (and they depend on the plugin, so it's not enough to "learn them once").
In order to figure out what they mean, a new user has to go through the reports of the individual calculations.
### Describe the solution you'd like
It would be great if a user could request to print the message associated with the exit status as a last column in `verdi process list`.
@sphuber What do you think?
| process | allow user to display meaning of exit status in verdi process list is your feature request related to a problem please describe verdi process list presents information about why calculations fail verdi process list a l pwcalculation pk created process label process state process status ago pwcalculation ☠ killed killed by parent ago pwcalculation ⏹ finished ago pwcalculation ☠ killed killed by parent ago pwcalculation ☠ killed killed by parent ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished some of it is self explanatory killed but the exit codes are not and they depend on the plugin so it s not enough to learn them once in order to figure out what they mean a new user has to go through the reports of the individual calculations describe the solution you d like it would be great if a user could request to print the message associated with the exit status as a last column in verdi process list sphuber what do you think | 1 |
20,357 | 27,014,590,628 | IssuesEvent | 2023-02-10 18:08:47 | googleapis/repo-automation-bots | https://api.github.com/repos/googleapis/repo-automation-bots | closed | Warning: a recent release failed | type: process | The following release PRs may have failed:
* #1464 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1414 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1366 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1064 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #998 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #992 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #984 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. | 1.0 | Warning: a recent release failed - The following release PRs may have failed:
* #1464 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1414 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1366 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #1064 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #998 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #992 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
* #984 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. | process | warning a recent release failed the following release prs may have failed the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published | 1 |
3,481 | 6,553,545,824 | IssuesEvent | 2017-09-05 23:18:14 | w3c/w3process | https://api.github.com/repos/w3c/w3process | closed | ToC/document mismatches | Editorial improvements Process2018Candidate | Section in the expanded table of contents, but not in the document:
* 6.2.2.1 Substantive Change
Sections in the document, but not in the expanded table of contents:
* 2.1.2.1 Membership Consortia
* 2.1.2.2 Related Members
* 2.1.3.1 Advisory Committee Mailing Lists
* 2.1.3.2 Advisory Committee Meetings
* 5.2.1.1 Member Representative in a Working Group
* 5.2.1.2 Member Representative in an Interest Group
* 5.2.1.3 Invited Expert in a Working Group
* 5.2.1.4 Invited Expert in an Interest Group
* 5.2.1.5 Team Representative in a Working Group
* 5.2.1.6 Team Representative in an Interest Group
Title mismatches between ToC and document:
* 6.2.5 "Classes of Changes to a Recommendation" vs. "Classes of Changes"
* 10.4 "Rejection of a Submission Request" vs. "Rejection of a Submission Request, and Submission Appeals"
Reference mismatch:
In [6.7.1 Errata Management](https://www.w3.org/2017/Process-20170301/#errata), there's a reference to "7.2.5 Classes of Changes", but it should be "6.2.5 Classes of Changes".
Version checked: Process Document 1 March 2017
(I haven't checked the latest Editor's version, but I think it would suffer from at least some of the problems above.)
Automatically numbered and generated ToC/references would be great.
| 1.0 | ToC/document mismatches - Section in the expanded table of contents, but not in the document:
* 6.2.2.1 Substantive Change
Sections in the document, but not in the expanded table of contents:
* 2.1.2.1 Membership Consortia
* 2.1.2.2 Related Members
* 2.1.3.1 Advisory Committee Mailing Lists
* 2.1.3.2 Advisory Committee Meetings
* 5.2.1.1 Member Representative in a Working Group
* 5.2.1.2 Member Representative in an Interest Group
* 5.2.1.3 Invited Expert in a Working Group
* 5.2.1.4 Invited Expert in an Interest Group
* 5.2.1.5 Team Representative in a Working Group
* 5.2.1.6 Team Representative in an Interest Group
Title mismatches between ToC and document:
* 6.2.5 "Classes of Changes to a Recommendation" vs. "Classes of Changes"
* 10.4 "Rejection of a Submission Request" vs. "Rejection of a Submission Request, and Submission Appeals"
Reference mismatch:
In [6.7.1 Errata Management](https://www.w3.org/2017/Process-20170301/#errata), there's a reference to "7.2.5 Classes of Changes", but it should be "6.2.5 Classes of Changes".
Version checked: Process Document 1 March 2017
(I haven't checked the latest Editor's version, but I think it would suffer from at least some of the problems above.)
Automatically numbered and generated ToC/references would be great.
| process | toc document mismatches section in the expanded table of contents but not in the document substantive change sections in the document but not in the expanded table of contents membership consortia related members advisory committee mailing lists advisory committee meetings member representative in a working group member representative in an interest group invited expert in a working group invited expert in an interest group team representative in a working group team representative in an interest group title mismatches between toc and document classes of changes to a recommendation vs classes of changes rejection of a submission request vs rejection of a submission request and submission appeals reference mismatch in there s a reference to classes of changes but it should be classes of changes version checked process document march i haven t checked the latest editor s version but i think it would suffer from at least some of the problems above automatically numbered and generated toc references would be great | 1 |
21,532 | 3,517,357,176 | IssuesEvent | 2016-01-12 06:59:43 | sanderv32/virtualboxserverservice | https://api.github.com/repos/sanderv32/virtualboxserverservice | closed | Installer hangs when trying to re-install | Component-Scripts Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Run the installer once (successfully)
2. Run it again (try to reinstall)
=> Installer hangs on "Creating scheduled task". Can't press cancel or even the
window.
I tried this on 64bit Windows 7 Ultimate with
VBoxService-setup_2.2.0.39135-BETA.exe.
```
Original issue reported on code.google.com by `jarno.el...@gmail.com` on 4 Dec 2010 at 4:34 | 1.0 | Installer hangs when trying to re-install - ```
What steps will reproduce the problem?
1. Run the installer once (successfully)
2. Run it again (try to reinstall)
=> Installer hangs on "Creating scheduled task". Can't press cancel or even the
window.
I tried this on 64bit Windows 7 Ultimate with
VBoxService-setup_2.2.0.39135-BETA.exe.
```
Original issue reported on code.google.com by `jarno.el...@gmail.com` on 4 Dec 2010 at 4:34 | non_process | installer hangs when trying to re install what steps will reproduce the problem run the installer once successfully run it again try to reinstall installer hangs on creating scheduled task can t press cancel or even the window i tried this on windows ultimate with vboxservice setup beta exe original issue reported on code google com by jarno el gmail com on dec at | 0 |
102,527 | 22,032,848,378 | IssuesEvent | 2022-05-28 05:21:45 | NicoVIII/Andromeda-for-GOG | https://api.github.com/repos/NicoVIII/Andromeda-for-GOG | closed | Add CodeTour | code | With the CodeTour extension for VScode it is possible to record Tours with comments and explain the structure of the project.
https://marketplace.visualstudio.com/items?itemName=vsls-contrib.codetour
This is a nice approach to make code easier to understand for new developers, so I want to try that. | 1.0 | Add CodeTour - With the CodeTour extension for VScode it is possible to record Tours with comments and explain the structure of the project.
https://marketplace.visualstudio.com/items?itemName=vsls-contrib.codetour
This is a nice approach to make code easier to understand for new developers, so I want to try that. | non_process | add codetour with the codetour extension for vscode it is possible to record tours with comments and explain the structure of the project this is a nice approach to make code easier to understand for new developers so i want to try that | 0 |
34,449 | 12,288,116,359 | IssuesEvent | 2020-05-09 15:20:00 | Zymergen/hubot-docker | https://api.github.com/repos/Zymergen/hubot-docker | opened | WS-2018-0209 (Medium) detected in morgan-1.6.1.tgz | security vulnerability | ## WS-2018-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/hubot-docker/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/hubot-docker/node_modules/morgan/package.json</p>
<p>
Dependency Hierarchy:
- hubot-2.19.0.tgz (Root Library)
- express-3.21.2.tgz
- connect-2.30.2.tgz
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Zymergen/hubot-docker/commit/07953cb6bb385a84410fb77bc2c3d2ff16dee495">07953cb6bb385a84410fb77bc2c3d2ff16dee495</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.
<p>Publish Date: 2018-11-25
<p>URL: <a href=>WS-2018-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","isTransitiveDependency":true,"dependencyTree":"hubot:2.19.0;express:3.21.2;connect:2.30.2;morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | True | WS-2018-0209 (Medium) detected in morgan-1.6.1.tgz - ## WS-2018-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/hubot-docker/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/hubot-docker/node_modules/morgan/package.json</p>
<p>
Dependency Hierarchy:
- hubot-2.19.0.tgz (Root Library)
- express-3.21.2.tgz
- connect-2.30.2.tgz
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Zymergen/hubot-docker/commit/07953cb6bb385a84410fb77bc2c3d2ff16dee495">07953cb6bb385a84410fb77bc2c3d2ff16dee495</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.
<p>Publish Date: 2018-11-25
<p>URL: <a href=>WS-2018-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","isTransitiveDependency":true,"dependencyTree":"hubot:2.19.0;express:3.21.2;connect:2.30.2;morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | non_process | ws medium detected in morgan tgz ws medium severity vulnerability vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file tmp ws scm hubot docker package json path to vulnerable library tmp ws scm hubot docker node modules morgan package json dependency hierarchy hubot tgz root library express tgz connect tgz x morgan tgz vulnerable library found in head commit a href vulnerability details morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack publish date url ws cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack vulnerabilityurl medium extradata | 0 |
15,682 | 19,847,795,320 | IssuesEvent | 2022-01-21 08:53:12 | ooi-data/RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample | https://api.github.com/repos/ooi-data/RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample | opened | 🛑 Processing failed: ValueError | process | ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T08:53:11.809404.
## Details
Flow name: `RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| 1.0 | 🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T08:53:11.809404.
## Details
Flow name: `RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| process | 🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name streamed ctdpf optode sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got | 1 |
2,217 | 2,599,855,153 | IssuesEvent | 2015-02-23 12:25:05 | lsecities/lsecities-wp-theme | https://api.github.com/repos/lsecities/lsecities-wp-theme | opened | DBUA Award lists - update template | enhancement needs:design | we should display some brief blurb for each award. photos also need to be larger than the current miserly thumbnails... | 1.0 | DBUA Award lists - update template - we should display some brief blurb for each award. photos also need to be larger than the current miserly thumbnails... | non_process | dbua award lists update template we should display some brief blurb for each award photos also need to be larger than the current miserly thumbnails | 0 |
19,796 | 26,178,434,019 | IssuesEvent | 2023-01-02 12:44:07 | prisma/prisma | https://api.github.com/repos/prisma/prisma | opened | Wrong db pull warning with multiSchema preview feature, it says that "These enums were enriched with `@@map` information taken from the previous Prisma schema." but the schema was actually empty. | bug/2-confirmed kind/bug process/candidate topic: introspection-warning topic: re-introspection team/schema topic: multiSchema | ### Bug description
The following warnings are confusing
```
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
```
Here we are in a re-introspection case but there is no datamodel, only a generator and a datasource.
The warning suggests that some information was taken from this empty datamodel which is impossible.
It's probably the logic in re-introspection in the engine that needs to be changed.
### How to reproduce
- create a database with `CREATE DATABASE "reintroMultiSchema";`
- import / execute the following SQL
```sql
-- CreateSchema
CREATE SCHEMA IF NOT EXISTS "base";
-- CreateSchema
CREATE SCHEMA IF NOT EXISTS "transactional";
-- CreateTable
CREATE TABLE "base"."User" (
"id" TEXT NOT NULL,
"email" TEXT NOT NULL,
CONSTRAINT "User_pkey" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "transactional"."Post" (
"id" TEXT NOT NULL,
"title" TEXT NOT NULL,
"authorId" TEXT NOT NULL,
CONSTRAINT "Post_pkey" PRIMARY KEY ("id")
);
-- AddForeignKey
ALTER TABLE "transactional"."Post" ADD CONSTRAINT "Post_authorId_fkey" FOREIGN KEY ("authorId") REFERENCES "base"."User"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
CREATE TYPE "base"."status" as enum ('ON','OFF');
--
-- tables names are renamed so they are identical
-- Original names are User and Post
-- Errors with Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field.
-- https://github.com/prisma/prisma/issues/15800
--
-- CreateTable
CREATE TABLE "base"."some_table" (
"id" TEXT NOT NULL,
"email" TEXT NOT NULL,
CONSTRAINT "User_pkey2" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "transactional"."some_table" (
"id" TEXT NOT NULL,
"title" TEXT NOT NULL,
"authorId" TEXT NOT NULL,
CONSTRAINT "Post_pkey2" PRIMARY KEY ("id")
);
-- AddForeignKey
ALTER TABLE "transactional"."some_table" ADD CONSTRAINT "Post_authorId_fkey2" FOREIGN KEY ("authorId") REFERENCES "base"."some_table"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
CREATE TYPE "transactional"."status" as enum ('ON','OFF');
```
- run `DATABASE_URL="postgres://prisma:prisma@localhost:5432/reintroMultiSchema" npx prisma@dev db pull --print --schema schema.prisma`
Result
```
generator client {
provider = "prisma-client-js"
previewFeatures = ["multiSchema"]
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
schemas = ["base", "transactional"]
}
model User {
id String @id
email String
Post Post[]
@@schema("base")
}
model base_some_table {
id String @id(map: "User_pkey2")
email String
some_table transactional_some_table[]
@@map("some_table")
@@schema("base")
}
model Post {
id String @id
title String
authorId String
User User @relation(fields: [authorId], references: [id])
@@schema("transactional")
}
model transactional_some_table {
id String @id(map: "Post_pkey2")
title String
authorId String
some_table base_some_table @relation(fields: [authorId], references: [id], map: "Post_authorId_fkey2")
@@map("some_table")
@@schema("transactional")
}
enum base_status {
ON
OFF
@@map("status")
@@schema("base")
}
enum transactional_status {
ON
OFF
@@map("status")
@@schema("transactional")
}
// introspectionSchemaVersion: NonPrisma
// *** WARNING ***
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
//
// These models and enums were renamed due to their names being duplicates in the Prisma Schema Language.
// Code 20
// [
// {
// "type": "Enum",
// "name": "base_status"
// },
// {
// "type": "Enum",
// "name": "transactional_status"
// },
// {
// "type": "Model",
// "name": "base_some_table"
// },
// {
// "type": "Model",
// "name": "transactional_some_table"
// }
// ]
//
```
### Expected behavior
These warnings should not be outputted
```
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
```
### Prisma information
```prisma
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
schemas = ["base", "transactional"]
}
generator client {
provider = "prisma-client-js"
previewFeatures = ["multiSchema"]
}
```
### Environment & setup
- OS: macOS
- Database: PostgreSQL
- Node.js version: NA
### Prisma Version
```
prisma : 4.9.0-dev.13
@prisma/client : Not found
Current platform : darwin-arm64
Query Engine (Node-API) : libquery-engine c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node)
Migration Engine : migration-engine-cli c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/migration-engine-darwin-arm64)
Introspection Engine : introspection-core c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/introspection-engine-darwin-arm64)
Format Binary : prisma-fmt c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/prisma-fmt-darwin-arm64)
Format Wasm : @prisma/prisma-fmt-wasm 4.9.0-6.c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35
Default Engines Hash : c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35
Studio : 0.479.0
Preview Features : multiSchema
```
| 1.0 | Wrong db pull warning with multiSchema preview feature, it says that "These enums were enriched with `@@map` information taken from the previous Prisma schema." but the schema was actually empty. - ### Bug description
The following warnings are confusing
```
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
```
Here we are in a re-introspection case but there is no datamodel, only a generator and a datasource.
The warning suggests that some information was taken from this empty datamodel which is impossible.
It's probably the logic in re-introspection in the engine that needs to be changed.
### How to reproduce
- create a database with `CREATE DATABASE "reintroMultiSchema";`
- import / execute the following SQL
```sql
-- CreateSchema
CREATE SCHEMA IF NOT EXISTS "base";
-- CreateSchema
CREATE SCHEMA IF NOT EXISTS "transactional";
-- CreateTable
CREATE TABLE "base"."User" (
"id" TEXT NOT NULL,
"email" TEXT NOT NULL,
CONSTRAINT "User_pkey" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "transactional"."Post" (
"id" TEXT NOT NULL,
"title" TEXT NOT NULL,
"authorId" TEXT NOT NULL,
CONSTRAINT "Post_pkey" PRIMARY KEY ("id")
);
-- AddForeignKey
ALTER TABLE "transactional"."Post" ADD CONSTRAINT "Post_authorId_fkey" FOREIGN KEY ("authorId") REFERENCES "base"."User"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
CREATE TYPE "base"."status" as enum ('ON','OFF');
--
-- tables names are renamed so they are identical
-- Original names are User and Post
-- Errors with Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field.
-- https://github.com/prisma/prisma/issues/15800
--
-- CreateTable
CREATE TABLE "base"."some_table" (
"id" TEXT NOT NULL,
"email" TEXT NOT NULL,
CONSTRAINT "User_pkey2" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "transactional"."some_table" (
"id" TEXT NOT NULL,
"title" TEXT NOT NULL,
"authorId" TEXT NOT NULL,
CONSTRAINT "Post_pkey2" PRIMARY KEY ("id")
);
-- AddForeignKey
ALTER TABLE "transactional"."some_table" ADD CONSTRAINT "Post_authorId_fkey2" FOREIGN KEY ("authorId") REFERENCES "base"."some_table"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
CREATE TYPE "transactional"."status" as enum ('ON','OFF');
```
- run `DATABASE_URL="postgres://prisma:prisma@localhost:5432/reintroMultiSchema" npx prisma@dev db pull --print --schema schema.prisma`
Result
```
generator client {
provider = "prisma-client-js"
previewFeatures = ["multiSchema"]
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
schemas = ["base", "transactional"]
}
model User {
id String @id
email String
Post Post[]
@@schema("base")
}
model base_some_table {
id String @id(map: "User_pkey2")
email String
some_table transactional_some_table[]
@@map("some_table")
@@schema("base")
}
model Post {
id String @id
title String
authorId String
User User @relation(fields: [authorId], references: [id])
@@schema("transactional")
}
model transactional_some_table {
id String @id(map: "Post_pkey2")
title String
authorId String
some_table base_some_table @relation(fields: [authorId], references: [id], map: "Post_authorId_fkey2")
@@map("some_table")
@@schema("transactional")
}
enum base_status {
ON
OFF
@@map("status")
@@schema("base")
}
enum transactional_status {
ON
OFF
@@map("status")
@@schema("transactional")
}
// introspectionSchemaVersion: NonPrisma
// *** WARNING ***
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
//
// These models and enums were renamed due to their names being duplicates in the Prisma Schema Language.
// Code 20
// [
// {
// "type": "Enum",
// "name": "base_status"
// },
// {
// "type": "Enum",
// "name": "transactional_status"
// },
// {
// "type": "Model",
// "name": "base_some_table"
// },
// {
// "type": "Model",
// "name": "transactional_some_table"
// }
// ]
//
```
### Expected behavior
These warnings should not be outputted
```
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "base_status"
//
// These enums were enriched with `@@map` information taken from the previous Prisma schema.
// - Enum "transactional_status"
```
### Prisma information
```prisma
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
schemas = ["base", "transactional"]
}
generator client {
provider = "prisma-client-js"
previewFeatures = ["multiSchema"]
}
```
### Environment & setup
- OS: macOS
- Database: PostgreSQL
- Node.js version: NA
### Prisma Version
```
prisma : 4.9.0-dev.13
@prisma/client : Not found
Current platform : darwin-arm64
Query Engine (Node-API) : libquery-engine c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node)
Migration Engine : migration-engine-cli c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/migration-engine-darwin-arm64)
Introspection Engine : introspection-core c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/introspection-engine-darwin-arm64)
Format Binary : prisma-fmt c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/prisma-fmt-darwin-arm64)
Format Wasm : @prisma/prisma-fmt-wasm 4.9.0-6.c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35
Default Engines Hash : c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35
Studio : 0.479.0
Preview Features : multiSchema
```
| process | wrong db pull warning with multischema preview feature it says that these enums were enriched with map information taken from the previous prisma schema but the schema was actually empty bug description the following warnings are confusing these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status here we are in a re introspection case but there is no datamodel only a generator and a datasource the warning suggests that some information was taken from this empty datamodel which is impossible it s probably the logic in re introspection in the engine that needs to be changed how to reproduce create a database with create database reintromultischema import execute the following sql sql createschema create schema if not exists base createschema create schema if not exists transactional createtable create table base user id text not null email text not null constraint user pkey primary key id createtable create table transactional post id text not null title text not null authorid text not null constraint post pkey primary key id addforeignkey alter table transactional post add constraint post authorid fkey foreign key authorid references base user id on delete restrict on update cascade create type base status as enum on off tables names are renamed so they are identical original names are user and post errors with error every relationinfo should have a complementary relationinfo on the opposite relation field createtable create table base some table id text not null email text not null constraint user primary key id createtable create table transactional some table id text not null title text not null authorid text not null constraint post primary key id addforeignkey alter table transactional some table add constraint post authorid foreign key authorid references base some table id on delete restrict on update cascade create type transactional status as enum on off run database url postgres prisma prisma localhost reintromultischema npx prisma dev db pull print schema schema prisma result generator client provider prisma client js previewfeatures datasource db provider postgresql url env database url schemas model user id string id email string post post schema base model base some table id string id map user email string some table transactional some table map some table schema base model post id string id title string authorid string user user relation fields references schema transactional model transactional some table id string id map post title string authorid string some table base some table relation fields references map post authorid map some table schema transactional enum base status on off map status schema base enum transactional status on off map status schema transactional introspectionschemaversion nonprisma warning these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status these models and enums were renamed due to their names being duplicates in the prisma schema language code type enum name base status type enum name transactional status type model name base some table type model name transactional some table expected behavior these warnings should not be outputted these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status prisma information prisma datasource db provider postgresql url env database url schemas generator client provider prisma client js previewfeatures environment setup os macos database postgresql node js version na prisma version prisma dev prisma client not found current platform darwin query engine node api libquery engine at npm npx node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at npm npx node modules prisma engines migration engine darwin introspection engine introspection core at npm npx node modules prisma engines introspection engine darwin format binary prisma fmt at npm npx node modules prisma engines prisma fmt darwin format wasm prisma prisma fmt wasm default engines hash studio preview features multischema | 1 |
386,956 | 11,453,716,869 | IssuesEvent | 2020-02-06 15:51:36 | easystats/parameters | https://api.github.com/repos/easystats/parameters | closed | Next CRAN release | high priority :running_man: | Due to changes in the latest R devel I have to update the performance and sjstats packages. however since parameters has new functions that are required for the latest performance function I think of submitting parameters first.
so are there any urgent features or bugs that needs to be fixed before submission? I'm planning to submit next week. | 1.0 | Next CRAN release - Due to changes in the latest R devel I have to update the performance and sjstats packages. however since parameters has new functions that are required for the latest performance function I think of submitting parameters first.
so are there any urgent features or bugs that needs to be fixed before submission? I'm planning to submit next week. | non_process | next cran release due to changes in the latest r devel i have to update the performance and sjstats packages however since parameters has new functions that are required for the latest performance function i think of submitting parameters first so are there any urgent features or bugs that needs to be fixed before submission i m planning to submit next week | 0 |
134,395 | 10,906,262,527 | IssuesEvent | 2019-11-20 12:34:22 | Laravel-Backpack/CRUD | https://api.github.com/repos/Laravel-Backpack/CRUD | closed | [4.0.7] select2_ajax filter breaks "Remove filters" button | testing or needs confirmation working on it | # Bug report
### What I did
1 - Add a select2_ajax filter
2 - Filter the results
3 - Click in the "Remove filters" button
### What I expected to happen
The filters should be removed
### What happened
The filter is removed from the interface, but the list return zero results:

### What I've already tried to fix it
What I can tell is that the filter key is still present when we remove the filter, but it has no value.

To avoid the empty results I added a check in the filter function:

### Backpack, Laravel, PHP, DB version
### PHP VERSION:
PHP 7.2.22 (cli) (built: Sep 8 2019 15:19:13) ( NTS )
### LARAVEL VERSION:
laravel/framework v6.1.0
### BACKPACK VERSION:
backpack/crud 4.0.7
backpack/generators 2.0.4
backpack/permissionmanager 5.0.1
| 1.0 | [4.0.7] select2_ajax filter breaks "Remove filters" button - # Bug report
### What I did
1 - Add a select2_ajax filter
2 - Filter the results
3 - Click in the "Remove filters" button
### What I expected to happen
The filters should be removed
### What happened
The filter is removed from the interface, but the list return zero results:

### What I've already tried to fix it
What I can tell is that the filter key is still present when we remove the filter, but it has no value.

To avoid the empty results I added a check in the filter function:

### Backpack, Laravel, PHP, DB version
### PHP VERSION:
PHP 7.2.22 (cli) (built: Sep 8 2019 15:19:13) ( NTS )
### LARAVEL VERSION:
laravel/framework v6.1.0
### BACKPACK VERSION:
backpack/crud 4.0.7
backpack/generators 2.0.4
backpack/permissionmanager 5.0.1
| non_process | ajax filter breaks remove filters button bug report what i did add a ajax filter filter the results click in the remove filters button what i expected to happen the filters should be removed what happened the filter is removed from the interface but the list return zero results what i ve already tried to fix it what i can tell is that the filter key is still present when we remove the filter but it has no value to avoid the empty results i added a check in the filter function backpack laravel php db version php version php cli built sep nts laravel version laravel framework backpack version backpack crud backpack generators backpack permissionmanager | 0 |
83,067 | 23,959,015,857 | IssuesEvent | 2022-09-12 17:19:11 | xamarin/xamarin-android | https://api.github.com/repos/xamarin/xamarin-android | closed | Generate OpenTK JCW just once and then merge them to all the API levels | Area: xamarin-android Build | Instead of generating them per level.
```
GenerateJavaCallableWrappers:
mono --debug=casts "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild/Xamarin/Android/jcw-gen.exe" -v10 -o "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/jcw/src" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0/Facades" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/OpenTK-1.0.dll"
"javac" -source 1.6 -target 1.6 -d "obj/Debug/jcw/bin" -bootclasspath "/Users/builder/android-toolchain/sdk/platforms/android-21/android.jar":"/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" @obj/Debug/jcw/classes.txt
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.
"jar" uf "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" -C "obj/Debug/jcw/bin" .
_GenerateMonoAndroidDex16:
"/Users/builder/android-toolchain/sdk/build-tools/28.0.0/dx" --dex --no-strict --output="/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.dex" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar"
```
That would mean to run `jcw-gen.exe` once, which will hopefully result in slightly shorter build times. | 1.0 | Generate OpenTK JCW just once and then merge them to all the API levels - Instead of generating them per level.
```
GenerateJavaCallableWrappers:
mono --debug=casts "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild/Xamarin/Android/jcw-gen.exe" -v10 -o "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/jcw/src" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0/Facades" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/OpenTK-1.0.dll"
"javac" -source 1.6 -target 1.6 -d "obj/Debug/jcw/bin" -bootclasspath "/Users/builder/android-toolchain/sdk/platforms/android-21/android.jar":"/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" @obj/Debug/jcw/classes.txt
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.
"jar" uf "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" -C "obj/Debug/jcw/bin" .
_GenerateMonoAndroidDex16:
"/Users/builder/android-toolchain/sdk/build-tools/28.0.0/dx" --dex --no-strict --output="/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.dex" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar"
```
That would mean to run `jcw-gen.exe` once, which will hopefully result in slightly shorter build times. | non_process | generate opentk jcw just once and then merge them to all the api levels instead of generating them per level generatejavacallablewrappers mono debug casts users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild xamarin android jcw gen exe o users builder jenkins workspace xamarin android pr builder xamarin android src opentk obj debug jcw src l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid facades users builder jenkins workspace xamarin android pr builder xamarin android src opentk obj debug opentk dll javac source target d obj debug jcw bin bootclasspath users builder android toolchain sdk platforms android android jar users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar obj debug jcw classes txt note some input files use unchecked or unsafe operations note recompile with xlint unchecked for details jar uf users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar c obj debug jcw bin users builder android toolchain sdk build tools dx dex no strict output users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android dex users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar that would mean to run jcw gen exe once which will hopefully result in slightly shorter build times | 0 |
13,361 | 15,825,459,797 | IssuesEvent | 2021-04-06 05:49:39 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | Add the ability to wire-up listeners before starting a child process | child_process feature request | **Is your feature request related to a problem? Please describe.**
When registering multiple listeners (callbacks) to the `data` event of a child process's `stdout`, there's no way to get the child process to wait for all the callbacks to be registered before starting. This means there's a window between registering the first listener and the second one in which the first listener might "pull" the first available chunk and when the second listener is registered, it won't receive the first chunk.
**A thinned-down example:**
```javascript
// Expected behaviour scenario
const { exec } = require("child_process");
p = exec("seq 1000"); // This command prints the numbers between 1 and 1000, each in a different line
a1 = '';
a2 = '';
p.stdout.on("data", (d) => a1 += d);
p.stdout.on("data", (d) => a2 += d);
// When the child process completes, a1 and a2 will both contain all the numbers from 1 to 1000
```
```javascript
// Edge-case scenario
const { exec } = require("child_process");
p = exec("seq 1000");
a1 = '';
a2 = '';
p.stdout.on("data", (d) => a1 += d);
setTimeout(() => {
p.stdout.on("data", (d) => a2 += d);
}, 500);
// When the child process completes, a1 will contain all the numbers from 1 to 1000 while a2 will remain an empty string
```
From what I understand from reading the documentation of child_process, when a child process is started nodejs saves it's output in a buffer until a listener is registered (either by directly binding to the 'data' event or by `pipe()`ing stdout to a writable stream). This behaviour creates two potential problems:
1. A second listener might not get the same data as the first one.
2. the child process might output more data than the buffer can contain before any data can be processed.
**Describe the solution you'd like**
The solution I propose is to allow wiring up all the listeners and pipes before starting the child process.
Considering backwards compatibility, I imagine the best way to achieve this is by passing a new option (something like `autostart` that will default to `true`) to the options parameter of `spawn`, `exec` etc., that will make those functions return a `ChildProcess` instance that was not yet started, together with a new `start()` method added to the `ChildProcess` class.
**Describe alternatives you've considered**
The alternatives as I see them are:
1. Only register a single handler and pass the data around to your multiple destinations.
2. Try to proxy the readable stream to a second one that is already wired up.
3. Try your best to minimize that window and hope for the best. | 1.0 | Add the ability to wire-up listeners before starting a child process - **Is your feature request related to a problem? Please describe.**
When registering multiple listeners (callbacks) to the `data` event of a child process's `stdout`, there's no way to get the child process to wait for all the callbacks to be registered before starting. This means there's a window between registering the first listener and the second one in which the first listener might "pull" the first available chunk and when the second listener is registered, it won't receive the first chunk.
**A thinned-down example:**
```javascript
// Expected behaviour scenario
const { exec } = require("child_process");
p = exec("seq 1000"); // This command prints the numbers between 1 and 1000, each in a different line
a1 = '';
a2 = '';
p.stdout.on("data", (d) => a1 += d);
p.stdout.on("data", (d) => a2 += d);
// When the child process completes, a1 and a2 will both contain all the numbers from 1 to 1000
```
```javascript
// Edge-case scenario
const { exec } = require("child_process");
p = exec("seq 1000");
a1 = '';
a2 = '';
p.stdout.on("data", (d) => a1 += d);
setTimeout(() => {
p.stdout.on("data", (d) => a2 += d);
}, 500);
// When the child process completes, a1 will contain all the numbers from 1 to 1000 while a2 will remain an empty string
```
From what I understand from reading the documentation of child_process, when a child process is started nodejs saves it's output in a buffer until a listener is registered (either by directly binding to the 'data' event or by `pipe()`ing stdout to a writable stream). This behaviour creates two potential problems:
1. A second listener might not get the same data as the first one.
2. the child process might output more data than the buffer can contain before any data can be processed.
**Describe the solution you'd like**
The solution I propose is to allow wiring up all the listeners and pipes before starting the child process.
Considering backwards compatibility, I imagine the best way to achieve this is by passing a new option (something like `autostart` that will default to `true`) to the options parameter of `spawn`, `exec` etc., that will make those functions return a `ChildProcess` instance that was not yet started, together with a new `start()` method added to the `ChildProcess` class.
**Describe alternatives you've considered**
The alternatives as I see them are:
1. Only register a single handler and pass the data around to your multiple destinations.
2. Try to proxy the readable stream to a second one that is already wired up.
3. Try your best to minimize that window and hope for the best. | process | add the ability to wire up listeners before starting a child process is your feature request related to a problem please describe when registering multiple listeners callbacks to the data event of a child process s stdout there s no way to get the child process to wait for all the callbacks to be registered before starting this means there s a window between registering the first listener and the second one in which the first listener might pull the first available chunk and when the second listener is registered it won t receive the first chunk a thinned down example javascript expected behaviour scenario const exec require child process p exec seq this command prints the numbers between and each in a different line p stdout on data d d p stdout on data d d when the child process completes and will both contain all the numbers from to javascript edge case scenario const exec require child process p exec seq p stdout on data d d settimeout p stdout on data d d when the child process completes will contain all the numbers from to while will remain an empty string from what i understand from reading the documentation of child process when a child process is started nodejs saves it s output in a buffer until a listener is registered either by directly binding to the data event or by pipe ing stdout to a writable stream this behaviour creates two potential problems a second listener might not get the same data as the first one the child process might output more data than the buffer can contain before any data can be processed describe the solution you d like the solution i propose is to allow wiring up all the listeners and pipes before starting the child process considering backwards compatibility i imagine the best way to achieve this is by passing a new option something like autostart that will default to true to the options parameter of spawn exec etc that will make those functions return a childprocess instance that was not yet started together with a new start method added to the childprocess class describe alternatives you ve considered the alternatives as i see them are only register a single handler and pass the data around to your multiple destinations try to proxy the readable stream to a second one that is already wired up try your best to minimize that window and hope for the best | 1 |
13,392 | 15,866,357,527 | IssuesEvent | 2021-04-08 15:39:53 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [Android] Study activities > Value picker response type >App is allowing the user to submit the response without selecting any value | Android Bug P1 Process: Fixed Process: Tested dev | Steps:-
1. Configure an activity with Value Picker based response type in Study builder
2. Open the Android application and login into App
3. Enroll into the study
4. Try to submit a response for Value Picker response type activities and verify
AR:- App is allowing the user to submit the response without selecting any value
ER:- App should not allow submitting the response until the user selects the value | 2.0 | [Android] Study activities > Value picker response type >App is allowing the user to submit the response without selecting any value - Steps:-
1. Configure an activity with Value Picker based response type in Study builder
2. Open the Android application and login into App
3. Enroll into the study
4. Try to submit a response for Value Picker response type activities and verify
AR:- App is allowing the user to submit the response without selecting any value
ER:- App should not allow submitting the response until the user selects the value | process | study activities value picker response type app is allowing the user to submit the response without selecting any value steps configure an activity with value picker based response type in study builder open the android application and login into app enroll into the study try to submit a response for value picker response type activities and verify ar app is allowing the user to submit the response without selecting any value er app should not allow submitting the response until the user selects the value | 1 |
6,708 | 9,815,605,176 | IssuesEvent | 2019-06-13 13:01:42 | AcademySoftwareFoundation/OpenCue | https://api.github.com/repos/AcademySoftwareFoundation/OpenCue | closed | Update project README to link to opencue.io | process | **Describe the bug**
When we deprecate the Wiki, we'll need to update the project README to link to opencue.io.
**To Reproduce**
N/A
**Expected behavior**
There should no longer be links from the main repository to the project Wiki.
**Screenshots**
N/A
**Version Number**
N/A
**Additional context**
N/A
| 1.0 | Update project README to link to opencue.io - **Describe the bug**
When we deprecate the Wiki, we'll need to update the project README to link to opencue.io.
**To Reproduce**
N/A
**Expected behavior**
There should no longer be links from the main repository to the project Wiki.
**Screenshots**
N/A
**Version Number**
N/A
**Additional context**
N/A
| process | update project readme to link to opencue io describe the bug when we deprecate the wiki we ll need to update the project readme to link to opencue io to reproduce n a expected behavior there should no longer be links from the main repository to the project wiki screenshots n a version number n a additional context n a | 1 |
25,761 | 12,734,975,061 | IssuesEvent | 2020-06-25 14:40:34 | kframework/kore | https://api.github.com/repos/kframework/kore | closed | Comparatively bad performance on some concrete tests | appetite: S investigate performance | When comparing the Haskell backend to the Java backend on KEVM concrete tests (574 tests total), we have:
- 392 tests which the Haskell backend finishes faster than the Java backend,
- 108 tests which finish within 2x the time of the Java backend (< 2x slowdown),
- 36 tests which timeout unexpectedly (see #1348 ), and
- 38 tests which take more than 2x the time of the Java backend.
This issue concerns the last category of tests, because it indicates that there are more efficient ways to handle those tests which the Java backend is taking advantage of but the Haskell backend is not. This is that list (note that we jump from 1.77x max slowdown in other tests to 2.82x slowdown in this list):
```
2.82 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_29.json
2.87 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_11.json
2.89 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_8.json
2.90 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_1.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_28.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_9.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/fibbonacci_unrolled.json
2.95 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_23.json
2.96 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_0.json
2.97 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_16.json
2.98 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_25.json
2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_31.json
2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_4.json
3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_21.json
3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_32.json
3.02 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_2.json
3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_10.json
3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_6.json
3.05 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_30.json
3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_12.json
3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_18.json
3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_14.json
3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_20.json
3.08 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_27.json
3.09 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_19.json
3.10 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_24.json
3.11 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_3.json
3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_17.json
3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_22.json
3.13 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_33.json
3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_26.json
3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_5.json
3.16 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_7.json
3.17 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_15.json
3.18 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_13.json
3.69 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpdestBigList.json
4.86 tests/ethereum-tests/VMTests/vmIOandFlowOperations/byte1.json
5.37 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpi_at_the_end.json
``` | True | Comparatively bad performance on some concrete tests - When comparing the Haskell backend to the Java backend on KEVM concrete tests (574 tests total), we have:
- 392 tests which the Haskell backend finishes faster than the Java backend,
- 108 tests which finish within 2x the time of the Java backend (< 2x slowdown),
- 36 tests which timeout unexpectedly (see #1348 ), and
- 38 tests which take more than 2x the time of the Java backend.
This issue concerns the last category of tests, because it indicates that there are more efficient ways to handle those tests which the Java backend is taking advantage of but the Haskell backend is not. This is that list (note that we jump from 1.77x max slowdown in other tests to 2.82x slowdown in this list):
```
2.82 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_29.json
2.87 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_11.json
2.89 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_8.json
2.90 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_1.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_28.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_9.json
2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/fibbonacci_unrolled.json
2.95 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_23.json
2.96 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_0.json
2.97 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_16.json
2.98 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_25.json
2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_31.json
2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_4.json
3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_21.json
3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_32.json
3.02 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_2.json
3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_10.json
3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_6.json
3.05 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_30.json
3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_12.json
3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_18.json
3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_14.json
3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_20.json
3.08 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_27.json
3.09 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_19.json
3.10 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_24.json
3.11 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_3.json
3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_17.json
3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_22.json
3.13 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_33.json
3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_26.json
3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_5.json
3.16 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_7.json
3.17 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_15.json
3.18 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_13.json
3.69 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpdestBigList.json
4.86 tests/ethereum-tests/VMTests/vmIOandFlowOperations/byte1.json
5.37 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpi_at_the_end.json
``` | non_process | comparatively bad performance on some concrete tests when comparing the haskell backend to the java backend on kevm concrete tests tests total we have tests which the haskell backend finishes faster than the java backend tests which finish within the time of the java backend slowdown tests which timeout unexpectedly see and tests which take more than the time of the java backend this issue concerns the last category of tests because it indicates that there are more efficient ways to handle those tests which the java backend is taking advantage of but the haskell backend is not this is that list note that we jump from max slowdown in other tests to slowdown in this list tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest fibbonacci unrolled json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmioandflowoperations jumpdestbiglist json tests ethereum tests vmtests vmioandflowoperations json tests ethereum tests vmtests vmioandflowoperations jumpi at the end json | 0 |
145,816 | 11,709,059,966 | IssuesEvent | 2020-03-08 16:43:44 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Reflect scan state with the right color in the UI | [zube]: To Test area/scan-tool team/ui | UI counterpart for https://github.com/rancher/rancher/issues/25214
UI needs to show "Fail" in Red and "Pass" in Green | 1.0 | Reflect scan state with the right color in the UI - UI counterpart for https://github.com/rancher/rancher/issues/25214
UI needs to show "Fail" in Red and "Pass" in Green | non_process | reflect scan state with the right color in the ui ui counterpart for ui needs to show fail in red and pass in green | 0 |
20,482 | 27,140,584,979 | IssuesEvent | 2023-02-16 16:04:44 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | closed | Add ignoreList to redaction processor | enhancement processor/redaction needs triage | ### Component(s)
processor/redaction
### Is your feature request related to a problem? Please describe.
Some fields may have a 12 digit number but it's an internal system identifier which doesn't benefit from redaction.
It would be nice if these attributes could be passed through without having the value treated as a credit card number.
### Describe the solution you'd like
Add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data.
### Describe alternatives you've considered
Other processors don't seem to have the regex matching and replacement ability that the redaction processor has.
### Additional context
The "allowList" in the processor seems to provide this capabillity at first glance but it is actually different. Adding the ignoreList will give users a way to proceed without having the modify their system identifiers or other drastic measures. | 1.0 | Add ignoreList to redaction processor - ### Component(s)
processor/redaction
### Is your feature request related to a problem? Please describe.
Some fields may have a 12 digit number but it's an internal system identifier which doesn't benefit from redaction.
It would be nice if these attributes could be passed through without having the value treated as a credit card number.
### Describe the solution you'd like
Add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data.
### Describe alternatives you've considered
Other processors don't seem to have the regex matching and replacement ability that the redaction processor has.
### Additional context
The "allowList" in the processor seems to provide this capabillity at first glance but it is actually different. Adding the ignoreList will give users a way to proceed without having the modify their system identifiers or other drastic measures. | process | add ignorelist to redaction processor component s processor redaction is your feature request related to a problem please describe some fields may have a digit number but it s an internal system identifier which doesn t benefit from redaction it would be nice if these attributes could be passed through without having the value treated as a credit card number describe the solution you d like add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data describe alternatives you ve considered other processors don t seem to have the regex matching and replacement ability that the redaction processor has additional context the allowlist in the processor seems to provide this capabillity at first glance but it is actually different adding the ignorelist will give users a way to proceed without having the modify their system identifiers or other drastic measures | 1 |
8,721 | 11,859,815,994 | IssuesEvent | 2020-03-25 13:58:55 | prisma/prisma2 | https://api.github.com/repos/prisma/prisma2 | opened | We should highlight that using prisma.raw() with parameters is not secure and recommend using prisma.raw`` | kind/discussion kind/docs kind/improvement process/candidate topic: prisma-client | ## Problem
Users are using `prisma.raw()` like
```js
const data = await prisma.raw(
`SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;`,
);
```
This example is using `prisma.raw()` the pure text version so there is no security around parameters.
Only raw`` is secure because it's using https://github.com/blakeembrey/sql-template-tag
## Solution
In this case it would be recommended to do
```js
const data = await prisma.raw`
SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;
`;
```
This should be highlighted in the docs (and examples?)
We also can think about how to warn users that are using `prisma.raw()` or even disable it under a flag?
*Note prisma.raw`` parameters do not work as of today see https://github.com/prisma/prisma-client-js/issues/595*
| 1.0 | We should highlight that using prisma.raw() with parameters is not secure and recommend using prisma.raw`` - ## Problem
Users are using `prisma.raw()` like
```js
const data = await prisma.raw(
`SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;`,
);
```
This example is using `prisma.raw()` the pure text version so there is no security around parameters.
Only raw`` is secure because it's using https://github.com/blakeembrey/sql-template-tag
## Solution
In this case it would be recommended to do
```js
const data = await prisma.raw`
SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;
`;
```
This should be highlighted in the docs (and examples?)
We also can think about how to warn users that are using `prisma.raw()` or even disable it under a flag?
*Note prisma.raw`` parameters do not work as of today see https://github.com/prisma/prisma-client-js/issues/595*
| process | we should highlight that using prisma raw with parameters is not secure and recommend using prisma raw problem users are using prisma raw like js const data await prisma raw select from provideritemattribute where provider item root id and user auth user id limit this example is using prisma raw the pure text version so there is no security around parameters only raw is secure because it s using solution in this case it would be recommended to do js const data await prisma raw select from provideritemattribute where provider item root id and user auth user id limit this should be highlighted in the docs and examples we also can think about how to warn users that are using prisma raw or even disable it under a flag note prisma raw parameters do not work as of today see | 1 |
86,705 | 10,788,967,321 | IssuesEvent | 2019-11-05 10:51:30 | buildsofwarcraft/app | https://api.github.com/repos/buildsofwarcraft/app | opened | Approval from Overwolf | design overwolf | - [ ] ask for approval from Overwolf regarding first design draft
- [ ] implement feedback from Overwolf if necessary | 1.0 | Approval from Overwolf - - [ ] ask for approval from Overwolf regarding first design draft
- [ ] implement feedback from Overwolf if necessary | non_process | approval from overwolf ask for approval from overwolf regarding first design draft implement feedback from overwolf if necessary | 0 |
19,825 | 26,214,749,921 | IssuesEvent | 2023-01-04 10:01:35 | qgis/QGIS-Documentation | https://api.github.com/repos/qgis/QGIS-Documentation | closed | Add relations export option to package alg (Request in QGIS) | Easy fix Processing Alg 3.28 | ### Request for documentation
From pull request QGIS/qgis#49520
Author: @elpaso
QGIS version: 3.28
**Add relations export option to package alg**
### PR Description:
Adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream, taking selected features into account.

Funded by: ARPA Piemonte
### Commits tagged with [need-docs] or [FEATURE] | 1.0 | Add relations export option to package alg (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#49520
Author: @elpaso
QGIS version: 3.28
**Add relations export option to package alg**
### PR Description:
Adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream, taking selected features into account.

Funded by: ARPA Piemonte
### Commits tagged with [need-docs] or [FEATURE] | process | add relations export option to package alg request in qgis request for documentation from pull request qgis qgis author elpaso qgis version add relations export option to package alg pr description adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream taking selected features into account funded by arpa piemonte commits tagged with or | 1 |
19,626 | 10,475,840,969 | IssuesEvent | 2019-09-23 17:14:03 | chapel-lang/chapel | https://api.github.com/repos/chapel-lang/chapel | closed | Slow Sparse Matrix Addition | area: Libraries / Modules type: Performance user issue | I tried using `.plus()` to add a large (square 10^6 with 10^6 nonzeros) matrix to it's transpose using
`var ppt = P.plus(PT);`
and the computation hadn't finished after 45 minutes so I killed it.
I wrote my own matrix addition method to work around this as follows
````
proc matPlus(A:[],B:[]) {
var dom = {A.domain.dim(1),B.domain.dim(2)};
var sps = CSRDomain(dom);
sps += A.domain;
sps += B.domain;
var S: [sps] real;
for (i,j) in sps {
S(i,j) = A(i,j) + B(i,j);
}
return S;
}
````
and with that instead, the same line ran in 26 seconds. | True | Slow Sparse Matrix Addition - I tried using `.plus()` to add a large (square 10^6 with 10^6 nonzeros) matrix to it's transpose using
`var ppt = P.plus(PT);`
and the computation hadn't finished after 45 minutes so I killed it.
I wrote my own matrix addition method to work around this as follows
````
proc matPlus(A:[],B:[]) {
var dom = {A.domain.dim(1),B.domain.dim(2)};
var sps = CSRDomain(dom);
sps += A.domain;
sps += B.domain;
var S: [sps] real;
for (i,j) in sps {
S(i,j) = A(i,j) + B(i,j);
}
return S;
}
````
and with that instead, the same line ran in 26 seconds. | non_process | slow sparse matrix addition i tried using plus to add a large square with nonzeros matrix to it s transpose using var ppt p plus pt and the computation hadn t finished after minutes so i killed it i wrote my own matrix addition method to work around this as follows proc matplus a b var dom a domain dim b domain dim var sps csrdomain dom sps a domain sps b domain var s real for i j in sps s i j a i j b i j return s and with that instead the same line ran in seconds | 0 |
155,927 | 5,962,703,358 | IssuesEvent | 2017-05-30 00:13:33 | input-output-hk/cardano-sl | https://api.github.com/repos/input-output-hk/cardano-sl | closed | [CSL-148] Setup a Hydra instance as CI | Mirroring Priority:Normal State:Done Type:Task | <blockquote>@domen</blockquote>
- Clean deploy (1TB EBS GP2, R3 instance) using nixops
- Copy existing ssh keys
- Slack integration
- Declarative jobsets
<!--MIRROR_META={"service":"youtrack","id":"CSL-148"}--> | 1.0 | [CSL-148] Setup a Hydra instance as CI - <blockquote>@domen</blockquote>
- Clean deploy (1TB EBS GP2, R3 instance) using nixops
- Copy existing ssh keys
- Slack integration
- Declarative jobsets
<!--MIRROR_META={"service":"youtrack","id":"CSL-148"}--> | non_process | setup a hydra instance as ci domen clean deploy ebs instance using nixops copy existing ssh keys slack integration declarative jobsets | 0 |
9,030 | 12,129,318,523 | IssuesEvent | 2020-04-22 22:17:41 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | opened | Kubernetes: Monitor API equivalent | enhancement process rest | **Problem**
The monitor API tests the REST API by running queries against it and validating the results are within certain parameters. It then exposes those results via an API and a separate dashboard shows the results. The monitor has to be explicitly configured with the REST endpoints, which is not possible to know in a containerized environments where pods come and go (especially with HPA). It's also another component we'd have to containerize and maintain.
Monitor tests:
```
Successfully called transactions with limit params only
Successfully retrieved transactions from with 50 seconds ago
Successfully retrieved balance from with 1800 seconds ago
Successfully called transactions with time and limit params
Successfully retrieved single transactions by id
Successfully called balances with time and limit params
Successfully called balances and performed account check
Successfully called accounts with time and limit params
Successfully called accounts for single account
Successfully called accounts and performed account check
Successfully called balances and performed account check
Successfully called transactions with order params only
```
**Solution**
- Add metrics that represent the tests that make sense
- For sure, the `with X seconds ago` tests can be made as metrics in importer (might already have an equivalent)
- Figure out if we should continue invoking every API ever X seconds or if a health check endpoint and helm acceptance test is enough for the others
**Alternatives**
- Containerize the monitor UI/API and have it list endpoints from k8s APIs
**Additional Context**
| 1.0 | Kubernetes: Monitor API equivalent - **Problem**
The monitor API tests the REST API by running queries against it and validating the results are within certain parameters. It then exposes those results via an API and a separate dashboard shows the results. The monitor has to be explicitly configured with the REST endpoints, which is not possible to know in a containerized environments where pods come and go (especially with HPA). It's also another component we'd have to containerize and maintain.
Monitor tests:
```
Successfully called transactions with limit params only
Successfully retrieved transactions from with 50 seconds ago
Successfully retrieved balance from with 1800 seconds ago
Successfully called transactions with time and limit params
Successfully retrieved single transactions by id
Successfully called balances with time and limit params
Successfully called balances and performed account check
Successfully called accounts with time and limit params
Successfully called accounts for single account
Successfully called accounts and performed account check
Successfully called balances and performed account check
Successfully called transactions with order params only
```
**Solution**
- Add metrics that represent the tests that make sense
- For sure, the `with X seconds ago` tests can be made as metrics in importer (might already have an equivalent)
- Figure out if we should continue invoking every API ever X seconds or if a health check endpoint and helm acceptance test is enough for the others
**Alternatives**
- Containerize the monitor UI/API and have it list endpoints from k8s APIs
**Additional Context**
| process | kubernetes monitor api equivalent problem the monitor api tests the rest api by running queries against it and validating the results are within certain parameters it then exposes those results via an api and a separate dashboard shows the results the monitor has to be explicitly configured with the rest endpoints which is not possible to know in a containerized environments where pods come and go especially with hpa it s also another component we d have to containerize and maintain monitor tests successfully called transactions with limit params only successfully retrieved transactions from with seconds ago successfully retrieved balance from with seconds ago successfully called transactions with time and limit params successfully retrieved single transactions by id successfully called balances with time and limit params successfully called balances and performed account check successfully called accounts with time and limit params successfully called accounts for single account successfully called accounts and performed account check successfully called balances and performed account check successfully called transactions with order params only solution add metrics that represent the tests that make sense for sure the with x seconds ago tests can be made as metrics in importer might already have an equivalent figure out if we should continue invoking every api ever x seconds or if a health check endpoint and helm acceptance test is enough for the others alternatives containerize the monitor ui api and have it list endpoints from apis additional context | 1 |
223,855 | 24,752,782,168 | IssuesEvent | 2022-10-21 14:59:43 | turkdevops/bit | https://api.github.com/repos/turkdevops/bit | closed | CVE-2015-9251 (Low) detected in jquery-1.7.2.min.js, jquery-1.9.0.min.js - autoclosed | security vulnerability | ## CVE-2015-9251 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.9.0.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.9.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p>
<p>Path to dependency file: /node_modules/chai-string/test/index.html</p>
<p>Path to vulnerable library: /node_modules/chai-string/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.0.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/bit/commit/2d5ba447ad6b7168a8a152ec95ebf4c142d6441e">2d5ba447ad6b7168a8a152ec95ebf4c142d6441e</a></p>
<p>Found in base branch: <b>app-generator</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-9251 (Low) detected in jquery-1.7.2.min.js, jquery-1.9.0.min.js - autoclosed - ## CVE-2015-9251 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.9.0.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.9.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p>
<p>Path to dependency file: /node_modules/chai-string/test/index.html</p>
<p>Path to vulnerable library: /node_modules/chai-string/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.0.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/bit/commit/2d5ba447ad6b7168a8a152ec95ebf4c142d6441e">2d5ba447ad6b7168a8a152ec95ebf4c142d6441e</a></p>
<p>Found in base branch: <b>app-generator</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve low detected in jquery min js jquery min js autoclosed cve low severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file node modules jmespath index html path to vulnerable library node modules jmespath index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node modules chai string test index html path to vulnerable library node modules chai string test index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch app generator vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend | 0 |
27,585 | 11,516,145,089 | IssuesEvent | 2020-02-14 03:51:41 | MadeByEmil/wdhan-basic | https://api.github.com/repos/MadeByEmil/wdhan-basic | closed | CVE-2015-9251 (Medium) detected in jquery-1.10.2.min.js | security vulnerability | ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/wdhan-basic/_includes/head.html</p>
<p>Path to vulnerable library: /wdhan-basic/_includes/head.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MadeByEmil/wdhan-basic/commit/20319c55082efc29a29a8cb4e602f2bb13bcd75f">20319c55082efc29a29a8cb4e602f2bb13bcd75f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-9251 (Medium) detected in jquery-1.10.2.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/wdhan-basic/_includes/head.html</p>
<p>Path to vulnerable library: /wdhan-basic/_includes/head.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MadeByEmil/wdhan-basic/commit/20319c55082efc29a29a8cb4e602f2bb13bcd75f">20319c55082efc29a29a8cb4e602f2bb13bcd75f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm wdhan basic includes head html path to vulnerable library wdhan basic includes head html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
89,070 | 11,194,813,953 | IssuesEvent | 2020-01-03 03:04:46 | ChildMindInstitute/mindlogger-app | https://api.github.com/repos/ChildMindInstitute/mindlogger-app | closed | ML-573 ⁃ MindLogger log in asks for username and email | Android Eden2016 design frontend iOS | Only username works to log in, but box prompts for email
 | 1.0 | ML-573 ⁃ MindLogger log in asks for username and email - Only username works to log in, but box prompts for email
 | non_process | ml ⁃ mindlogger log in asks for username and email only username works to log in but box prompts for email | 0 |
131,177 | 18,214,827,622 | IssuesEvent | 2021-09-30 02:01:16 | samjcs/xcp-ng-org | https://api.github.com/repos/samjcs/xcp-ng-org | opened | CVE-2018-20676 (Medium) detected in bootstrap-3.3.5.min.js | security vulnerability | ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: xcp-ng-org/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>Path to vulnerable library: /node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samjcs/xcp-ng-org/commit/76bae0317ed38f295fb61df7a9af82f659a38c78">76bae0317ed38f295fb61df7a9af82f659a38c78</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.5","packageFilePaths":["/node_modules/autocomplete.js/test/playground_jquery.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-20676 (Medium) detected in bootstrap-3.3.5.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: xcp-ng-org/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>Path to vulnerable library: /node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samjcs/xcp-ng-org/commit/76bae0317ed38f295fb61df7a9af82f659a38c78">76bae0317ed38f295fb61df7a9af82f659a38c78</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.5","packageFilePaths":["/node_modules/autocomplete.js/test/playground_jquery.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_process | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file xcp ng org node modules autocomplete js test playground jquery html path to vulnerable library node modules autocomplete js test playground jquery html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree twitter bootstrap isminimumfixversionavailable true minimumfixversion bootstrap basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the tooltip data viewport attribute vulnerabilityurl | 0 |
113,845 | 24,499,057,831 | IssuesEvent | 2022-10-10 11:14:14 | assemblee-virtuelle/semapps | https://api.github.com/repos/assemblee-virtuelle/semapps | closed | Partager comment remonter des erreurs sur Sentry | 5. Qualité de code 11. Infoculteur low priority | https://sentry.io/welcome/
Fait par @srosset81 sur le serveur Colibris, à partager avec les autres.
Eventuellement le mettre en place sur Archipel ? | 1.0 | Partager comment remonter des erreurs sur Sentry - https://sentry.io/welcome/
Fait par @srosset81 sur le serveur Colibris, à partager avec les autres.
Eventuellement le mettre en place sur Archipel ? | non_process | partager comment remonter des erreurs sur sentry fait par sur le serveur colibris à partager avec les autres eventuellement le mettre en place sur archipel | 0 |
2,823 | 5,773,309,470 | IssuesEvent | 2017-04-28 01:27:06 | gaocegege/maintainer | https://api.github.com/repos/gaocegege/maintainer | opened | Detect the language for CONTRIBUTING | process/not claimed type/feature | It is the base of language specific CONTRIBUTING generation. | 1.0 | Detect the language for CONTRIBUTING - It is the base of language specific CONTRIBUTING generation. | process | detect the language for contributing it is the base of language specific contributing generation | 1 |
4,580 | 7,410,057,882 | IssuesEvent | 2018-03-21 00:01:01 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Video quality only low -> unreadeable | active-directory assigned-to-author docs-experience in-process triaged | Hi,
Not sure what's wrong, but the videoplayer informs me only "Low" quality is avaialble for the videos, so the screens are not readeable.
Valentijn
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 751889da-f001-117f-d525-3b27d467edf2
* Version Independent ID: ce46adec-de86-7c34-45ba-d735cba76d2d
* Content: [How do Azure Active Directory admins add B2B collaboration users?](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-b2b-admin-add-users#feedback)
* Content Source: [articles/active-directory/active-directory-b2b-admin-add-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-b2b-admin-add-users.md)
* Service: **active-directory**
* GitHub Login: @twooley
* Microsoft Alias: **twooley** | 1.0 | Video quality only low -> unreadeable - Hi,
Not sure what's wrong, but the videoplayer informs me only "Low" quality is avaialble for the videos, so the screens are not readeable.
Valentijn
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 751889da-f001-117f-d525-3b27d467edf2
* Version Independent ID: ce46adec-de86-7c34-45ba-d735cba76d2d
* Content: [How do Azure Active Directory admins add B2B collaboration users?](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-b2b-admin-add-users#feedback)
* Content Source: [articles/active-directory/active-directory-b2b-admin-add-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-b2b-admin-add-users.md)
* Service: **active-directory**
* GitHub Login: @twooley
* Microsoft Alias: **twooley** | process | video quality only low unreadeable hi not sure what s wrong but the videoplayer informs me only low quality is avaialble for the videos so the screens are not readeable valentijn document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service active directory github login twooley microsoft alias twooley | 1 |
323,174 | 23,937,399,790 | IssuesEvent | 2022-09-11 12:33:25 | os-climate/ITR | https://api.github.com/repos/os-climate/ITR | opened | Units: mboe vs mmboe | documentation | We all know that "boe" is a "Barrel of Oil Equivalent".
It appears that the 'm' in "mboe" is a Roman M, meaning thousand, and that 'mm' is a Roman thousand squared, meaning a million.
Is that correct? Is 'mcf' therefore a thousand cu ft of gas? Are there other 'm's that are thousands instead of millions we need to look out for?
I will need to update our units files (which don't know this particular prefixing rule). | 1.0 | Units: mboe vs mmboe - We all know that "boe" is a "Barrel of Oil Equivalent".
It appears that the 'm' in "mboe" is a Roman M, meaning thousand, and that 'mm' is a Roman thousand squared, meaning a million.
Is that correct? Is 'mcf' therefore a thousand cu ft of gas? Are there other 'm's that are thousands instead of millions we need to look out for?
I will need to update our units files (which don't know this particular prefixing rule). | non_process | units mboe vs mmboe we all know that boe is a barrel of oil equivalent it appears that the m in mboe is a roman m meaning thousand and that mm is a roman thousand squared meaning a million is that correct is mcf therefore a thousand cu ft of gas are there other m s that are thousands instead of millions we need to look out for i will need to update our units files which don t know this particular prefixing rule | 0 |
192,269 | 15,342,345,966 | IssuesEvent | 2021-02-27 15:51:44 | jsinger0420/NodeEra | https://api.github.com/repos/jsinger0420/NodeEra | opened | Python / QT development environment | documentation | Need to write instructions on how to install and run the application from python. | 1.0 | Python / QT development environment - Need to write instructions on how to install and run the application from python. | non_process | python qt development environment need to write instructions on how to install and run the application from python | 0 |
15,730 | 10,265,612,819 | IssuesEvent | 2019-08-22 19:17:41 | ualbertalib/avalon | https://api.github.com/repos/ualbertalib/avalon | closed | Move Publish/unpublish, Edit and Delete buttons to the right side of objects and make more prominent | Post-launch usability | ### Descriptive summary
It is easy for users to miss where the publish, edit and delete buttons are
### Expected behavior
Publish, edit and delete buttons are right justified on object views
Buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme
### Actual behavior
Publish, edit and delete buttons are left justified and easy to miss
### Steps to reproduce the behavior
1. Go to any record view and observe the placement and style of these buttons | True | Move Publish/unpublish, Edit and Delete buttons to the right side of objects and make more prominent - ### Descriptive summary
It is easy for users to miss where the publish, edit and delete buttons are
### Expected behavior
Publish, edit and delete buttons are right justified on object views
Buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme
### Actual behavior
Publish, edit and delete buttons are left justified and easy to miss
### Steps to reproduce the behavior
1. Go to any record view and observe the placement and style of these buttons | non_process | move publish unpublish edit and delete buttons to the right side of objects and make more prominent descriptive summary it is easy for users to miss where the publish edit and delete buttons are expected behavior publish edit and delete buttons are right justified on object views buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme actual behavior publish edit and delete buttons are left justified and easy to miss steps to reproduce the behavior go to any record view and observe the placement and style of these buttons | 0 |
5,344 | 8,176,038,838 | IssuesEvent | 2018-08-28 05:48:39 | TEAMMATES/teammates | https://api.github.com/repos/TEAMMATES/teammates | closed | Update workflow/process documents | a-Process c.DevOps | There has been some internal restructuring as well as improvements to GitHub that necessitates updates to the workflow/process document. To name a few:
- No longer need to ask dev to update branch if outdated since GitHub provides ["Allow edits from maintainers"](https://github.com/blog/2247-improving-collaboration-with-forks) (since very long ago)
- [Multiple issue/PR templates](https://github.com/blog/2495-multiple-issue-and-pull-request-templates)
- [Label descriptions](https://github.com/blog/2505-label-improvements-emoji-descriptions-and-more) (already adopted but the docs need update)
- Removing local test
- Removing new contributor intro
- Renaming `s.ToMerge` -> `s.FinalReview` and `s.MergeApproved` -> `s.ToMerge`
Also, the core team structure has changed.
Will take some time this weekend to update. | 1.0 | Update workflow/process documents - There has been some internal restructuring as well as improvements to GitHub that necessitates updates to the workflow/process document. To name a few:
- No longer need to ask dev to update branch if outdated since GitHub provides ["Allow edits from maintainers"](https://github.com/blog/2247-improving-collaboration-with-forks) (since very long ago)
- [Multiple issue/PR templates](https://github.com/blog/2495-multiple-issue-and-pull-request-templates)
- [Label descriptions](https://github.com/blog/2505-label-improvements-emoji-descriptions-and-more) (already adopted but the docs need update)
- Removing local test
- Removing new contributor intro
- Renaming `s.ToMerge` -> `s.FinalReview` and `s.MergeApproved` -> `s.ToMerge`
Also, the core team structure has changed.
Will take some time this weekend to update. | process | update workflow process documents there has been some internal restructuring as well as improvements to github that necessitates updates to the workflow process document to name a few no longer need to ask dev to update branch if outdated since github provides since very long ago already adopted but the docs need update removing local test removing new contributor intro renaming s tomerge s finalreview and s mergeapproved s tomerge also the core team structure has changed will take some time this weekend to update | 1 |
7,907 | 11,089,904,242 | IssuesEvent | 2019-12-14 22:01:00 | dita-ot/dita-ot | https://api.github.com/repos/dita-ot/dita-ot | closed | topicpull stylesheet error | bug preprocess preprocess2 stale | Hi,
I got the stylesheet error when running my test data.
```
topicpull:
[topicpull] Transforming into D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp
[topicpull] Loading stylesheet D:\DITA-OT\dita-ot-2.4.6\xsl\preprocess\topicpull.xsl
[topicpull] Processing D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp\topics-en\p_programmingelements.xml
[topicpull] Error at xsl:param on line 623 of topicpullImpl.xsl:
[topicpull] XTDE0610: A value must be supplied for the parameter because the default value is not a
[topicpull] valid instance of the required type
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#96)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#55)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]
[topicpull] in built-in template rule
[topicpull] Failed to transform document: A value must be supplied for the parameter because the default value is not a valid instance of the required type
```
The test data is here:
[20170419-sample-en-for-word.zip](https://github.com/dita-ot/dita-ot/files/1017504/20170419-sample-en-for-word.zip)
The command-line is here:
```
PS D:\DITA-OT\dita-ot-2.4.6> bin\dita -i "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\sample_en.ditamap" -f pdf2 -o "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\out\pdf2" -l "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\log.txt" -v -t "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp" --pdf.formatter=ah
Unable to locate tools.jar. Expected to find it in C:\Program Files\Java\jre1.8.0_131\lib\tools.jar
PS D:\DITA-OT\dita-ot-2.4.6>
```
As this data was written in the past, If it contains authoring mistake, please let me know:
Regards,
--
/*--------------------------------------------------
Toshihiko Makita
Development Group. Antenna House, Inc. Ina Branch
Web site:
http://www.antenna.co.jp/
http://www.antennahouse.com/
--------------------------------------------------*/
| 2.0 | topicpull stylesheet error - Hi,
I got the stylesheet error when running my test data.
```
topicpull:
[topicpull] Transforming into D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp
[topicpull] Loading stylesheet D:\DITA-OT\dita-ot-2.4.6\xsl\preprocess\topicpull.xsl
[topicpull] Processing D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp\topics-en\p_programmingelements.xml
[topicpull] Error at xsl:param on line 623 of topicpullImpl.xsl:
[topicpull] XTDE0610: A value must be supplied for the parameter because the default value is not a
[topicpull] valid instance of the required type
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#96)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#55)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]/section[8]
[topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161)
[topicpull] processing /topic/body[1]
[topicpull] in built-in template rule
[topicpull] Failed to transform document: A value must be supplied for the parameter because the default value is not a valid instance of the required type
```
The test data is here:
[20170419-sample-en-for-word.zip](https://github.com/dita-ot/dita-ot/files/1017504/20170419-sample-en-for-word.zip)
The command-line is here:
```
PS D:\DITA-OT\dita-ot-2.4.6> bin\dita -i "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\sample_en.ditamap" -f pdf2 -o "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\out\pdf2" -l "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\log.txt" -v -t "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp" --pdf.formatter=ah
Unable to locate tools.jar. Expected to find it in C:\Program Files\Java\jre1.8.0_131\lib\tools.jar
PS D:\DITA-OT\dita-ot-2.4.6>
```
As this data was written in the past, If it contains authoring mistake, please let me know:
Regards,
--
/*--------------------------------------------------
Toshihiko Makita
Development Group. Antenna House, Inc. Ina Branch
Web site:
http://www.antenna.co.jp/
http://www.antennahouse.com/
--------------------------------------------------*/
| process | topicpull stylesheet error hi i got the stylesheet error when running my test data topicpull transforming into d svn testdata sample en for word temp loading stylesheet d dita ot dita ot xsl preprocess topicpull xsl processing d svn testdata sample en for word temp topics en p programmingelements xml error at xsl param on line of topicpullimpl xsl a value must be supplied for the parameter because the default value is not a valid instance of the required type at xsl apply templates file d dita ot dita ot xsl preprocess topicpull pr d xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpull pr d xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section syntaxdiagram at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body in built in template rule failed to transform document a value must be supplied for the parameter because the default value is not a valid instance of the required type the test data is here the command line is here ps d dita ot dita ot bin dita i d svn testdata sample en for word sample en ditamap f o d svn testdata sample en for word out l d svn testdata sample en for word log txt v t d svn testdata sample en for word temp pdf formatter ah unable to locate tools jar expected to find it in c program files java lib tools jar ps d dita ot dita ot as this data was written in the past if it contains authoring mistake please let me know regards toshihiko makita development group antenna house inc ina branch web site | 1 |
28,817 | 13,836,059,873 | IssuesEvent | 2020-10-14 00:07:12 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | CustomScrollView + SliverAppBar + WebView | P3 a: annoyance a: platform-views cp: 1.22 cp: 1.22 completed engine found in release: 1.20 has reproducible steps p: first party p: webview plugin severe: performance waiting for PR to land (fixed) | Hi,
When trying to implement a floating `SliverAppBar` in conjunction with a `WebView` Flutter drops a significant amount of frames.
It seems like the WebView is resizing itself whenever the `SliverAppBar` is scrolled out of view.
```
@override
Widget build(BuildContext context) {
return CustomScrollView(
slivers: <Widget>[
SliverAppBar(
title: const Text("Heading"),
floating: true,
),
SliverFillRemaining(
child: WebView(initialUrl: "http://stackoverflow.com"),
)
],
);
}
```
As seen on this GIF:

| True | CustomScrollView + SliverAppBar + WebView - Hi,
When trying to implement a floating `SliverAppBar` in conjunction with a `WebView` Flutter drops a significant amount of frames.
It seems like the WebView is resizing itself whenever the `SliverAppBar` is scrolled out of view.
```
@override
Widget build(BuildContext context) {
return CustomScrollView(
slivers: <Widget>[
SliverAppBar(
title: const Text("Heading"),
floating: true,
),
SliverFillRemaining(
child: WebView(initialUrl: "http://stackoverflow.com"),
)
],
);
}
```
As seen on this GIF:

| non_process | customscrollview sliverappbar webview hi when trying to implement a floating sliverappbar in conjunction with a webview flutter drops a significant amount of frames it seems like the webview is resizing itself whenever the sliverappbar is scrolled out of view override widget build buildcontext context return customscrollview slivers sliverappbar title const text heading floating true sliverfillremaining child webview initialurl as seen on this gif | 0 |
11,959 | 14,726,198,674 | IssuesEvent | 2021-01-06 06:22:31 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | Test failure: System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows | area-System.Diagnostics.Process | failed in job: [runtime-libraries-coreclr outerloop 20210105.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=938368&view=ms.vss-test-web.build-test-results-tab&runId=29665462&resultId=101960&paneView=debug)
net6.0-windows-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c
Error message
~~~
System.UnauthorizedAccessException : Attempted to perform an unauthorized operation.
Stack trace
at System.Security.AccessControl.Win32.SetSecurityInfo(ResourceType type, String name, SafeHandle handle, SecurityInfos securityInformation, SecurityIdentifier owner, SecurityIdentifier group, GenericAcl sacl, GenericAcl dacl) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/Win32.cs:line 314
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, SafeHandle handle, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 263
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 353
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 343
at System.Security.AccessControl.FileSystemSecurity.Persist(String fullPath) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/Security/AccessControl/FileSystemSecurity.cs:line 124
at System.IO.FileSystemAclExtensions.SetAccessControl(FileInfo fileInfo, FileSecurity fileSecurity) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/IO/FileSystemAclExtensions.cs:line 78
at System.Diagnostics.Tests.ProcessStartInfoTests.SetAccessControl(String userName, String filePath, Boolean add) in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 534
at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 503
~~~ | 1.0 | Test failure: System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows - failed in job: [runtime-libraries-coreclr outerloop 20210105.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=938368&view=ms.vss-test-web.build-test-results-tab&runId=29665462&resultId=101960&paneView=debug)
net6.0-windows-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c
Error message
~~~
System.UnauthorizedAccessException : Attempted to perform an unauthorized operation.
Stack trace
at System.Security.AccessControl.Win32.SetSecurityInfo(ResourceType type, String name, SafeHandle handle, SecurityInfos securityInformation, SecurityIdentifier owner, SecurityIdentifier group, GenericAcl sacl, GenericAcl dacl) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/Win32.cs:line 314
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, SafeHandle handle, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 263
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 353
at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 343
at System.Security.AccessControl.FileSystemSecurity.Persist(String fullPath) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/Security/AccessControl/FileSystemSecurity.cs:line 124
at System.IO.FileSystemAclExtensions.SetAccessControl(FileInfo fileInfo, FileSecurity fileSecurity) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/IO/FileSystemAclExtensions.cs:line 78
at System.Diagnostics.Tests.ProcessStartInfoTests.SetAccessControl(String userName, String filePath, Boolean add) in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 534
at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 503
~~~ | process | test failure system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows failed in job windows release coreclr release windows server core open windows open mcr microsoft com dotnet buildtools prereqs windowsservercore helix error message system unauthorizedaccessexception attempted to perform an unauthorized operation stack trace at system security accesscontrol setsecurityinfo resourcetype type string name safehandle handle securityinfos securityinformation securityidentifier owner securityidentifier group genericacl sacl genericacl dacl in src libraries system security accesscontrol src system security accesscontrol cs line at system security accesscontrol nativeobjectsecurity persist string name safehandle handle accesscontrolsections includesections object exceptioncontext in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol nativeobjectsecurity persist string name accesscontrolsections includesections object exceptioncontext in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol nativeobjectsecurity persist string name accesscontrolsections includesections in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol filesystemsecurity persist string fullpath in src libraries system io filesystem accesscontrol src system security accesscontrol filesystemsecurity cs line at system io filesystemaclextensions setaccesscontrol fileinfo fileinfo filesecurity filesecurity in src libraries system io filesystem accesscontrol src system io filesystemaclextensions cs line at system diagnostics tests processstartinfotests setaccesscontrol string username string filepath boolean add in src libraries system diagnostics process tests processstartinfotests cs line at system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows in src libraries system diagnostics process tests processstartinfotests cs line | 1 |
7,998 | 11,188,267,117 | IssuesEvent | 2020-01-02 03:48:32 | AlmuraDev/SGCraft | https://api.github.com/repos/AlmuraDev/SGCraft | closed | Stargates doesn't deal enough damage | bug in process | When I wear a Draconic Evolution armor the event horizon doesn't even hurt me.
Could you increase it's damage to ensure that anything killable WILL die when they are at the wrong place at the wrong time? | 1.0 | Stargates doesn't deal enough damage - When I wear a Draconic Evolution armor the event horizon doesn't even hurt me.
Could you increase it's damage to ensure that anything killable WILL die when they are at the wrong place at the wrong time? | process | stargates doesn t deal enough damage when i wear a draconic evolution armor the event horizon doesn t even hurt me could you increase it s damage to ensure that anything killable will die when they are at the wrong place at the wrong time | 1 |
35,399 | 14,681,673,078 | IssuesEvent | 2020-12-31 14:00:24 | eventespresso/barista | https://api.github.com/repos/eventespresso/barista | closed | Investigate isRTL from i18n and check Next component | C: services 🤝 D: Packages 📦 P3: med priority 😐 T: bug 🐞 T: task 🧹 | Currently `isRTL()` is returning `false` when a RTL language is used in WP. The same result is obtained with `wp.i18n.isRTL()` | 1.0 | Investigate isRTL from i18n and check Next component - Currently `isRTL()` is returning `false` when a RTL language is used in WP. The same result is obtained with `wp.i18n.isRTL()` | non_process | investigate isrtl from and check next component currently isrtl is returning false when a rtl language is used in wp the same result is obtained with wp isrtl | 0 |
2,652 | 5,429,403,363 | IssuesEvent | 2017-03-03 18:22:08 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | Missing or truncated error message | confirmed-bug process | This looks close to #6456, and I thought it was just another case of that issue, but @indutny mentioned that the errors are printed from c++, so #6456 alone shouldn't have caused this, this is why I'm opening a separate issue.
A harder to reproduce example where the error is missing (it reproduces with about 10% chance for me):
``` js
for (var i = 0; i < 10000; i++) {
console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i);
}
a();
```
A simplier to reproduce example (the error gets truncated here most of the times, and sometimes is missing as in the previous example):
``` js
for (var i = 0; i < 10000; i++) {
console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i);
}
throw new Error(Array(100000 + 1).join('x') + '!');
```
| 1.0 | Missing or truncated error message - This looks close to #6456, and I thought it was just another case of that issue, but @indutny mentioned that the errors are printed from c++, so #6456 alone shouldn't have caused this, this is why I'm opening a separate issue.
A harder to reproduce example where the error is missing (it reproduces with about 10% chance for me):
``` js
for (var i = 0; i < 10000; i++) {
console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i);
}
a();
```
A simplier to reproduce example (the error gets truncated here most of the times, and sometimes is missing as in the previous example):
``` js
for (var i = 0; i < 10000; i++) {
console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i);
}
throw new Error(Array(100000 + 1).join('x') + '!');
```
| process | missing or truncated error message this looks close to and i thought it was just another case of that issue but indutny mentioned that the errors are printed from c so alone shouldn t have caused this this is why i m opening a separate issue a harder to reproduce example where the error is missing it reproduces with about chance for me js for var i i i console log hellohellohellohellohellohelellohellohellohellohellohellohello i a a simplier to reproduce example the error gets truncated here most of the times and sometimes is missing as in the previous example js for var i i i console log hellohellohellohellohellohelellohellohellohellohellohellohello i throw new error array join x | 1 |
59,224 | 14,369,086,827 | IssuesEvent | 2020-12-01 09:18:45 | ignatandrei/stankins | https://api.github.com/repos/ignatandrei/stankins | closed | CVE-2019-6284 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2019-6284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.10.0.tgz</b>, <b>node-sass-4.9.3.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.10.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.11.4.tgz (Root Library)
- :x: **node-sass-4.10.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-sass-4.9.3.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.10.7.tgz (Root Library)
- :x: **node-sass-4.9.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-6284 (Medium) detected in multiple libraries - ## CVE-2019-6284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.10.0.tgz</b>, <b>node-sass-4.9.3.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.10.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.11.4.tgz (Root Library)
- :x: **node-sass-4.10.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-sass-4.9.3.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.10.7.tgz (Root Library)
- :x: **node-sass-4.9.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm stankins solution stankinsdatawebangular package json path to vulnerable library tmp ws scm stankins solution stankinsdatawebangular node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm stankins solution stankinsaliveangular package json path to vulnerable library tmp ws scm stankins solution stankinsaliveangular node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a heap based buffer over read exists in sass prelexer alternatives in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
7,527 | 10,600,052,478 | IssuesEvent | 2019-10-10 09:16:25 | Altinn/altinn-studio | https://api.github.com/repos/Altinn/altinn-studio | closed | Analyse and define how app frontend and app backend should handle process | analysis api-expose app-backend app-frontend process | As part of #243 there is defined some new API's for handling of process.
There are some decisions that needs to be made on how the process api should work and how App Frontend should use the API
The below diagram is a suggestion and are used as input for defining the clarification needed

The flow above is
1. User clicks link to a App from service catalog
2. The browser redirect to the start service page in the app. The react application renders the correct view
3. User click on instantiate button.
4. App Frontend calls instansiate API
5. Instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to.
6. User/browser open the instance
7. App frontend calls app backed to get the instance state
8. App Frontend render the correct component for the given state.
9. User input data . App Frontend call Data API to update data. If calculation or dataretrieavel causes additional data updates on server side, the response indicates to app frontend to get the latest data
10. App frontend gets the updated data
9 and 10 is repeated as long user has data to input
11. User press complete/send inn button
12. App frontend Saves data
13. App frontend calls validation API and any possible errors are presented to the user
14. If no errors are returned App Frontend call Process API to complete Task. Process Api trigger task validation to make sure it is allowed to move process forward. If validation fails, process API returns
### Define interaction between API and consumers (App Frontend, end user systems, agency systems)
We need to define exactly how App Frontend and End User System will interact with process
### App logic for task completion
We need to define what kind of control the app developer should
Should we let the App developer add code to a HandleTaskCompletion
### Error message when task in process is tried to be completed
When App Frontend post to Process API to complete task it might be condition for the current task that is not valid. It could be validation errors. Should validation errors be return or not?
| 1.0 | Analyse and define how app frontend and app backend should handle process - As part of #243 there is defined some new API's for handling of process.
There are some decisions that needs to be made on how the process api should work and how App Frontend should use the API
The below diagram is a suggestion and are used as input for defining the clarification needed

The flow above is
1. User clicks link to a App from service catalog
2. The browser redirect to the start service page in the app. The react application renders the correct view
3. User click on instantiate button.
4. App Frontend calls instansiate API
5. Instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to.
6. User/browser open the instance
7. App frontend calls app backed to get the instance state
8. App Frontend render the correct component for the given state.
9. User input data . App Frontend call Data API to update data. If calculation or dataretrieavel causes additional data updates on server side, the response indicates to app frontend to get the latest data
10. App frontend gets the updated data
9 and 10 is repeated as long user has data to input
11. User press complete/send inn button
12. App frontend Saves data
13. App frontend calls validation API and any possible errors are presented to the user
14. If no errors are returned App Frontend call Process API to complete Task. Process Api trigger task validation to make sure it is allowed to move process forward. If validation fails, process API returns
### Define interaction between API and consumers (App Frontend, end user systems, agency systems)
We need to define exactly how App Frontend and End User System will interact with process
### App logic for task completion
We need to define what kind of control the app developer should
Should we let the App developer add code to a HandleTaskCompletion
### Error message when task in process is tried to be completed
When App Frontend post to Process API to complete task it might be condition for the current task that is not valid. It could be validation errors. Should validation errors be return or not?
| process | analyse and define how app frontend and app backend should handle process as part of there is defined some new api s for handling of process there are some decisions that needs to be made on how the process api should work and how app frontend should use the api the below diagram is a suggestion and are used as input for defining the clarification needed the flow above is user clicks link to a app from service catalog the browser redirect to the start service page in the app the react application renders the correct view user click on instantiate button app frontend calls instansiate api instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to user browser open the instance app frontend calls app backed to get the instance state app frontend render the correct component for the given state user input data app frontend call data api to update data if calculation or dataretrieavel causes additional data updates on server side the response indicates to app frontend to get the latest data app frontend gets the updated data and is repeated as long user has data to input user press complete send inn button app frontend saves data app frontend calls validation api and any possible errors are presented to the user if no errors are returned app frontend call process api to complete task process api trigger task validation to make sure it is allowed to move process forward if validation fails process api returns define interaction between api and consumers app frontend end user systems agency systems we need to define exactly how app frontend and end user system will interact with process app logic for task completion we need to define what kind of control the app developer should should we let the app developer add code to a handletaskcompletion error message when task in process is tried to be completed when app frontend post to process api to complete task it might be condition for the current task that is not valid it could be validation errors should validation errors be return or not | 1 |
6,222 | 9,160,890,597 | IssuesEvent | 2019-03-01 08:58:36 | fablabbcn/fablabs.io | https://api.github.com/repos/fablabbcn/fablabs.io | closed | Bug in the labs list when adding a new lab | Approval Process bug | When adding a new lab, in the labs list there are some strange items (more than one!):

| 1.0 | Bug in the labs list when adding a new lab - When adding a new lab, in the labs list there are some strange items (more than one!):

| process | bug in the labs list when adding a new lab when adding a new lab in the labs list there are some strange items more than one | 1 |
351,826 | 25,040,480,598 | IssuesEvent | 2022-11-04 20:10:15 | CMPUT301F22T15/Foodverse | https://api.github.com/repos/CMPUT301F22T15/Foodverse | closed | Update CRC Cards For Part 3 Submission | documentation | - [x] Ensure CRC cards are up to date with classes in the code, updating where needed.
- [x] #55 | 1.0 | Update CRC Cards For Part 3 Submission - - [x] Ensure CRC cards are up to date with classes in the code, updating where needed.
- [x] #55 | non_process | update crc cards for part submission ensure crc cards are up to date with classes in the code updating where needed | 0 |
11,965 | 14,729,400,203 | IssuesEvent | 2021-01-06 11:24:14 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | Mobile App > Modify study list fields | Android P1 Process: Enhancement Process: Track 3 iOS | Modify study list to remove fields in accordance with #2600 . Adjust the UI.
Have the study list display the default image provided by the Study Builder if there is no alternate image uploaded. | 2.0 | Mobile App > Modify study list fields - Modify study list to remove fields in accordance with #2600 . Adjust the UI.
Have the study list display the default image provided by the Study Builder if there is no alternate image uploaded. | process | mobile app modify study list fields modify study list to remove fields in accordance with adjust the ui have the study list display the default image provided by the study builder if there is no alternate image uploaded | 1 |
11,672 | 14,531,508,754 | IssuesEvent | 2020-12-14 20:54:05 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | "Import geotagged photos" fails for jpg / jpeg on Windows | Bug Processing Windows | <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
`Import geotagged photos` algorithm fails for jpg / jpeg on Windows.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
1. Unzip [sample_jpg.zip](https://github.com/qgis/QGIS/files/5688584/sample_jpg.zip)
2. Use the `Import geotagged photos` algorithm from the processing toolbox with the input folder containing the sample image.
4. See error --> `Could not open ...`
**QGIS and OS versions**
QGIS version | 3.16.1-Hannover | QGIS code revision | b381a90dca
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | Discovery; ImportPhotos-adapted_path; joinmultiplelines; mmqgis; QPackage; db_manager; MetaSearch; processing
**Additional context**
This happens for a lot of different images captured with different devices.
On Linux it works with QGIS 3.16.1, but it's compiled against GDAL/OGR **2.4.0**.
Might be related to GDAL but I'm not entirely sure (`gdalinfo` fails also: https://github.com/OSGeo/gdal/issues/3284).
Also with several image viewers (e.g. GIMP) the sample jpg works flawlessly and it's possible to show the EXIF data. | 1.0 | "Import geotagged photos" fails for jpg / jpeg on Windows - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
`Import geotagged photos` algorithm fails for jpg / jpeg on Windows.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
1. Unzip [sample_jpg.zip](https://github.com/qgis/QGIS/files/5688584/sample_jpg.zip)
2. Use the `Import geotagged photos` algorithm from the processing toolbox with the input folder containing the sample image.
4. See error --> `Could not open ...`
**QGIS and OS versions**
QGIS version | 3.16.1-Hannover | QGIS code revision | b381a90dca
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | Discovery; ImportPhotos-adapted_path; joinmultiplelines; mmqgis; QPackage; db_manager; MetaSearch; processing
**Additional context**
This happens for a lot of different images captured with different devices.
On Linux it works with QGIS 3.16.1, but it's compiled against GDAL/OGR **2.4.0**.
Might be related to GDAL but I'm not entirely sure (`gdalinfo` fails also: https://github.com/OSGeo/gdal/issues/3284).
Also with several image viewers (e.g. GIMP) the sample jpg works flawlessly and it's possible to show the EXIF data. | process | import geotagged photos fails for jpg jpeg on windows bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug import geotagged photos algorithm fails for jpg jpeg on windows how to reproduce unzip use the import geotagged photos algorithm from the processing toolbox with the input folder containing the sample image see error could not open qgis and os versions qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins discovery importphotos adapted path joinmultiplelines mmqgis qpackage db manager metasearch processing additional context this happens for a lot of different images captured with different devices on linux it works with qgis but it s compiled against gdal ogr might be related to gdal but i m not entirely sure gdalinfo fails also also with several image viewers e g gimp the sample jpg works flawlessly and it s possible to show the exif data | 1 |
107,047 | 13,430,853,091 | IssuesEvent | 2020-09-07 05:51:14 | ibmdb/go_ibm_db | https://api.github.com/repos/ibmdb/go_ibm_db | closed | New version does not work for parameterized queries | As Designed duplicate | We recently updated to new version 0.3.0, and existing queries are no longer working.
Query like: `SELECT * FROM dummy_table WHERE cod_1 = ? AND cod_2 = ? AND cod_3 = '24'`
Provides a **CLI0100E Wrong number of parameters. SQLSTATE=07001**
I've been debugging a bit, and seen that provided args are not used in https://github.com/ibmdb/go_ibm_db/blob/6d3b2e8acd84b85fba599aa27246921a9b5482d2/conn.go#L50
Using DB2Driver 11.5
Any idea with that? | 1.0 | New version does not work for parameterized queries - We recently updated to new version 0.3.0, and existing queries are no longer working.
Query like: `SELECT * FROM dummy_table WHERE cod_1 = ? AND cod_2 = ? AND cod_3 = '24'`
Provides a **CLI0100E Wrong number of parameters. SQLSTATE=07001**
I've been debugging a bit, and seen that provided args are not used in https://github.com/ibmdb/go_ibm_db/blob/6d3b2e8acd84b85fba599aa27246921a9b5482d2/conn.go#L50
Using DB2Driver 11.5
Any idea with that? | non_process | new version does not work for parameterized queries we recently updated to new version and existing queries are no longer working query like select from dummy table where cod and cod and cod provides a wrong number of parameters sqlstate i ve been debugging a bit and seen that provided args are not used in using any idea with that | 0 |
15,767 | 19,913,145,692 | IssuesEvent | 2022-01-25 19:21:06 | input-output-hk/high-assurance-legacy | https://api.github.com/repos/input-output-hk/high-assurance-legacy | closed | Automate the application of quasi-compatibility laws | type: enhancement reason: wontfix language: isabelle topic: process calculus | While in the ♮-calculus `▹` and `▹⇧∞` are compatible with bisimilarity, in the Þ-calculus they are not but only what we call quasi-compatible. While this still allows us to prove typical statements that we used to prove using compatibility, the proofs become quite verbose, since the lack of proper compatibility means that `process_family_equivalence` cannot rewrite under those combinators.
We shall add a proof method that allows at least a limited form of rewriting under combinators like `▹` and `▹⇧∞` based on quasi-compatibility. Since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities, we shall name it `bisimilarity`. We shall keep `process_family_equivalence`, as `bisimilarity` will most likely not work with processes that perform branching based on received values. | 1.0 | Automate the application of quasi-compatibility laws - While in the ♮-calculus `▹` and `▹⇧∞` are compatible with bisimilarity, in the Þ-calculus they are not but only what we call quasi-compatible. While this still allows us to prove typical statements that we used to prove using compatibility, the proofs become quite verbose, since the lack of proper compatibility means that `process_family_equivalence` cannot rewrite under those combinators.
We shall add a proof method that allows at least a limited form of rewriting under combinators like `▹` and `▹⇧∞` based on quasi-compatibility. Since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities, we shall name it `bisimilarity`. We shall keep `process_family_equivalence`, as `bisimilarity` will most likely not work with processes that perform branching based on received values. | process | automate the application of quasi compatibility laws while in the ♮ calculus ▹ and ▹⇧∞ are compatible with bisimilarity in the þ calculus they are not but only what we call quasi compatible while this still allows us to prove typical statements that we used to prove using compatibility the proofs become quite verbose since the lack of proper compatibility means that process family equivalence cannot rewrite under those combinators we shall add a proof method that allows at least a limited form of rewriting under combinators like ▹ and ▹⇧∞ based on quasi compatibility since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities we shall name it bisimilarity we shall keep process family equivalence as bisimilarity will most likely not work with processes that perform branching based on received values | 1 |
103,751 | 16,609,058,764 | IssuesEvent | 2021-06-02 09:14:50 | dreamboy9/ofbiz | https://api.github.com/repos/dreamboy9/ofbiz | opened | CVE-2020-7656 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.3.2.min.js</b>, <b>jquery-1.8.1.min.js</b>, <b>jquery-1.6.2.js</b></p></summary>
<p>
<details><summary><b>jquery-1.3.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_standalone.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.3.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.6.2.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/index.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.6.2.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/dreamboy9/ofbiz/commit/9f6ed39589395d00f1d69228cb50a7987ba11512">9f6ed39589395d00f1d69228cb50a7987ba11512</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7656 (Medium) detected in multiple libraries - ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.3.2.min.js</b>, <b>jquery-1.8.1.min.js</b>, <b>jquery-1.6.2.js</b></p></summary>
<p>
<details><summary><b>jquery-1.3.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_standalone.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.3.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.6.2.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js</a></p>
<p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/index.html</p>
<p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.6.2.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/dreamboy9/ofbiz/commit/9f6ed39589395d00f1d69228cb50a7987ba11512">9f6ed39589395d00f1d69228cb50a7987ba11512</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery js jquery min js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules underscore string test test standalone html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules underscore string test test underscore vendor jquery js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules tap node modules redeyed examples browser index html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules tap node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules qs test browser index html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules qs test browser jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch trunk vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
68,485 | 21,664,953,483 | IssuesEvent | 2022-05-07 03:07:52 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Timestamp's position on Message Edits history modal window | T-Defect S-Tolerable A-Message-Editing A-Appearance O-Uncommon | ### Steps to reproduce
1. Create a test room
2. Send a message
3. Edit the message
4. View the edit history
### Outcome
#### What did you expect?
The timestamp should be displayed on the left side.
#### What happened instead?
It is displayed on the center of the row. I guess the root cause is same as https://github.com/vector-im/element-web/issues/22011.

### Operating system
Debian
### Browser information
Firefox
### URL for webapp
localhost
### Application version
3e31fdb6a71f43774420e8da32452861296a263a
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Timestamp's position on Message Edits history modal window - ### Steps to reproduce
1. Create a test room
2. Send a message
3. Edit the message
4. View the edit history
### Outcome
#### What did you expect?
The timestamp should be displayed on the left side.
#### What happened instead?
It is displayed on the center of the row. I guess the root cause is same as https://github.com/vector-im/element-web/issues/22011.

### Operating system
Debian
### Browser information
Firefox
### URL for webapp
localhost
### Application version
3e31fdb6a71f43774420e8da32452861296a263a
### Homeserver
_No response_
### Will you send logs?
No | non_process | timestamp s position on message edits history modal window steps to reproduce create a test room send a message edit the message view the edit history outcome what did you expect the timestamp should be displayed on the left side what happened instead it is displayed on the center of the row i guess the root cause is same as operating system debian browser information firefox url for webapp localhost application version homeserver no response will you send logs no | 0 |
185,751 | 6,727,384,932 | IssuesEvent | 2017-10-17 13:27:57 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | TinyTILE bluetooth app flash | area: Boards bug priority: low | **_Reported by Jie Zhou:_**
Hello,
I'm switching boards from arduino 101 to tinyTILE, the flashing method for tinyTILE is different than the arduino. Below are the commands I used for arduino flashing.
```
$ 1 make BOARD=tinytile
$ 2 dfu-util -a x86_app -D outdir/arduino_101/zephyr.bin
```
I can make BOARD=tinyTILE. But when I enter in dfu-util -a x86_app -D outdir/tinytile/zephyr.bin the command line tells me outdir/tinytile/zephyr.bin cannot open file. Yet in outdir, after make BOARD command, tinytile directory is included.
I would like to flash tinyTILE using dfu. Does anyone know the exact command?
Thanks,
Jie
(Imported from Jira ZEP-2600) | 1.0 | TinyTILE bluetooth app flash - **_Reported by Jie Zhou:_**
Hello,
I'm switching boards from arduino 101 to tinyTILE, the flashing method for tinyTILE is different than the arduino. Below are the commands I used for arduino flashing.
```
$ 1 make BOARD=tinytile
$ 2 dfu-util -a x86_app -D outdir/arduino_101/zephyr.bin
```
I can make BOARD=tinyTILE. But when I enter in dfu-util -a x86_app -D outdir/tinytile/zephyr.bin the command line tells me outdir/tinytile/zephyr.bin cannot open file. Yet in outdir, after make BOARD command, tinytile directory is included.
I would like to flash tinyTILE using dfu. Does anyone know the exact command?
Thanks,
Jie
(Imported from Jira ZEP-2600) | non_process | tinytile bluetooth app flash reported by jie zhou hello i m switching boards from arduino to tinytile the flashing method for tinytile is different than the arduino below are the commands i used for arduino flashing make board tinytile dfu util a app d outdir arduino zephyr bin i can make board tinytile but when i enter in dfu util a app d outdir tinytile zephyr bin the command line tells me outdir tinytile zephyr bin cannot open file yet in outdir after make board command tinytile directory is included i would like to flash tinytile using dfu does anyone know the exact command thanks jie imported from jira zep | 0 |
46,053 | 5,996,412,311 | IssuesEvent | 2017-06-03 14:06:11 | raulir/bccms | https://api.github.com/repos/raulir/bccms | closed | toolbar - put less used buttons under dropdown button | cms design/UX enhancement/refactoring | toolbar - put less used buttons under dropdown button | 1.0 | toolbar - put less used buttons under dropdown button - toolbar - put less used buttons under dropdown button | non_process | toolbar put less used buttons under dropdown button toolbar put less used buttons under dropdown button | 0 |
20,934 | 27,779,912,770 | IssuesEvent | 2023-03-16 20:09:44 | cse442-at-ub/project_s23-cinco | https://api.github.com/repos/cse442-at-ub/project_s23-cinco | opened | Add functionality to profile, account settings, and login button on navbar | Processing Task Sprint 2 | *Task Tests*
run in "Sprint2-Navbar-Buttons" branch in github
test1:
- in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser
- look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"
-click on the "profile" button to get to edit profile page:
-click on the "account settings" to get to the account settings page:
-click on "login" to get to login page:
| 1.0 | Add functionality to profile, account settings, and login button on navbar - *Task Tests*
run in "Sprint2-Navbar-Buttons" branch in github
test1:
- in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser
- look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"
-click on the "profile" button to get to edit profile page:
-click on the "account settings" to get to the account settings page:
-click on "login" to get to login page:
| process | add functionality to profile account settings and login button on navbar task tests run in navbar buttons branch in github in the project folder project cinco run npm start in the terminal this will open the homepage in a browser look at the navbar and ensure it has the following buttons profile account settings login click on the profile button to get to edit profile page click on the account settings to get to the account settings page click on login to get to login page | 1 |
17,097 | 5,330,138,783 | IssuesEvent | 2017-02-15 16:24:38 | eclipse/che | https://api.github.com/repos/eclipse/che | closed | The "Projects" info page does not refresh after deleting a project. | kind/bug severity/P2 status/code-review | **Reproduction Steps:**
1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**.
2. Create any project.
3. Select the project and click "Delete" button .
4. Confirm deleting.
**Expected behavior:**
The project is removed and not visible in the "Projects" tab.
**Observed behavior:**
The project is removed but the projects list isn't refreshed. The deleted project name is still visible.
**Che version:** 5.3.0-SNAPSHOT
**OS and version:** Ubuntu 16.04 LTS
**Docker version:** 1.12.3
**Che install:** Docker container
**Additional information:**
Problem can be reliably reproduced, doesn't happen randomly: [Yes]

| 1.0 | The "Projects" info page does not refresh after deleting a project. - **Reproduction Steps:**
1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**.
2. Create any project.
3. Select the project and click "Delete" button .
4. Confirm deleting.
**Expected behavior:**
The project is removed and not visible in the "Projects" tab.
**Observed behavior:**
The project is removed but the projects list isn't refreshed. The deleted project name is still visible.
**Che version:** 5.3.0-SNAPSHOT
**OS and version:** Ubuntu 16.04 LTS
**Docker version:** 1.12.3
**Che install:** Docker container
**Additional information:**
Problem can be reliably reproduced, doesn't happen randomly: [Yes]

| non_process | the projects info page does not refresh after deleting a project reproduction steps workspaces configure workspace select tab projects create any project select the project and click delete button confirm deleting expected behavior the project is removed and not visible in the projects tab observed behavior the project is removed but the projects list isn t refreshed the deleted project name is still visible che version snapshot os and version ubuntu lts docker version che install docker container additional information problem can be reliably reproduced doesn t happen randomly | 0 |
4,398 | 7,294,514,949 | IssuesEvent | 2018-02-26 00:23:15 | nyu-software-engineering/online-time-tracker | https://api.github.com/repos/nyu-software-engineering/online-time-tracker | closed | Read Chrome extension documentation and create a trivial extension | 2-in process Sprint 0 task | This task is part of User Story #14 | 1.0 | Read Chrome extension documentation and create a trivial extension - This task is part of User Story #14 | process | read chrome extension documentation and create a trivial extension this task is part of user story | 1 |
710,754 | 24,432,293,097 | IssuesEvent | 2022-10-06 08:59:57 | unep-grid/map-x-mgl | https://api.github.com/repos/unep-grid/map-x-mgl | closed | Missing translations & typos [end 2021 update] | priority 1 text/Translation | _This issue is an update of #327_
### Edit style
- [x] change "opacity" to "transparency" to be consistent with the rest of the application
### Area intersection tools
- [x] _select analysis_ section is not translated
### Manage sources
- [x] pop-up title: Edit source -> Gérer la source / Editar la fuente de datos
### Search engine API
- [x] all terms in the panel
### Map composer
- [x] all terms in the module
### Draw tool
- [x] missing entry for `draw_feature_type`
### Story maps engine
- [x] slide -> In Spanish should be "Diapositiva"
- [x] slide classes (card, text-right ....)
- [x] in edition mode: all tooltips are not translated
### Attribut table
- [x] clear filter
- [x] export CSV
### Other remarks
- [x] ~texts accompanying the progress bars are not translated~ Will be refactored
- [x] ~logs are not translated~
- [x] lack of consistency between the terms Configure and Edit e.g. the "Configure view" button becomes "Edit view (in the header). | 1.0 | Missing translations & typos [end 2021 update] - _This issue is an update of #327_
### Edit style
- [x] change "opacity" to "transparency" to be consistent with the rest of the application
### Area intersection tools
- [x] _select analysis_ section is not translated
### Manage sources
- [x] pop-up title: Edit source -> Gérer la source / Editar la fuente de datos
### Search engine API
- [x] all terms in the panel
### Map composer
- [x] all terms in the module
### Draw tool
- [x] missing entry for `draw_feature_type`
### Story maps engine
- [x] slide -> In Spanish should be "Diapositiva"
- [x] slide classes (card, text-right ....)
- [x] in edition mode: all tooltips are not translated
### Attribut table
- [x] clear filter
- [x] export CSV
### Other remarks
- [x] ~texts accompanying the progress bars are not translated~ Will be refactored
- [x] ~logs are not translated~
- [x] lack of consistency between the terms Configure and Edit e.g. the "Configure view" button becomes "Edit view (in the header). | non_process | missing translations typos this issue is an update of edit style change opacity to transparency to be consistent with the rest of the application area intersection tools select analysis section is not translated manage sources pop up title edit source gérer la source editar la fuente de datos search engine api all terms in the panel map composer all terms in the module draw tool missing entry for draw feature type story maps engine slide in spanish should be diapositiva slide classes card text right in edition mode all tooltips are not translated attribut table clear filter export csv other remarks texts accompanying the progress bars are not translated will be refactored logs are not translated lack of consistency between the terms configure and edit e g the configure view button becomes edit view in the header | 0 |
14,689 | 17,836,486,711 | IssuesEvent | 2021-09-03 02:19:57 | jim-king-2000/IndustryCamera | https://api.github.com/repos/jim-king-2000/IndustryCamera | closed | [bug]: 多路拉流时,停止一路流,所有流都会停止 | bug processing A | ### 问题描述
多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断
### 您预期的行为
停止一路拉流,不影响其他端拉流
### 系统表现的行为
停止一路拉流,其他端拉流终端
### 复现路径
开启两个网页端拉流,停止一路,会发现另一路流中断
### 辅助信息
- 浏览器版本:Edge/Chrome 92
- 固件版本:v1.0 | 1.0 | [bug]: 多路拉流时,停止一路流,所有流都会停止 - ### 问题描述
多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断
### 您预期的行为
停止一路拉流,不影响其他端拉流
### 系统表现的行为
停止一路拉流,其他端拉流终端
### 复现路径
开启两个网页端拉流,停止一路,会发现另一路流中断
### 辅助信息
- 浏览器版本:Edge/Chrome 92
- 固件版本:v1.0 | process | 多路拉流时,停止一路流,所有流都会停止 问题描述 多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断 您预期的行为 停止一路拉流,不影响其他端拉流 系统表现的行为 停止一路拉流,其他端拉流终端 复现路径 开启两个网页端拉流,停止一路,会发现另一路流中断 辅助信息 浏览器版本:edge chrome 固件版本: | 1 |
15,788 | 19,977,800,686 | IssuesEvent | 2022-01-29 11:35:34 | bdrum/kaggle | https://api.github.com/repos/bdrum/kaggle | opened | Create pipeline for titanic | enhancement preprocessing model classification pipeline | I have to prepare pipeline for a data that will include such steps as
- [ ] feature selection
- [ ] preprocessing
- [ ] classification | 1.0 | Create pipeline for titanic - I have to prepare pipeline for a data that will include such steps as
- [ ] feature selection
- [ ] preprocessing
- [ ] classification | process | create pipeline for titanic i have to prepare pipeline for a data that will include such steps as feature selection preprocessing classification | 1 |
11,901 | 14,697,307,547 | IssuesEvent | 2021-01-04 02:50:53 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Case sensitive field calculator (using graphical modeler) | Feature Request Modeller Processing | **Feature description.**
At the moment, Field Calculator does not seem to be case sensitive when creating a new field. Therefore, it is not possible to simply rename a field for which you can't decide a name (e.g. `Distance` to `distance` created with `Distance matrix`)
It would be great to be able to do such action without having to overburden the model (see below).
**Additional context**
There are some alternatives, as discussed on this SO post : [https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler](https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler)
Yet, using the `Refactoring fields` algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler.
Currently, the best alternative is to add several steps to the process : create a temporary new field, copy the desired value in this field, then rename the temporary field.
| 1.0 | Case sensitive field calculator (using graphical modeler) - **Feature description.**
At the moment, Field Calculator does not seem to be case sensitive when creating a new field. Therefore, it is not possible to simply rename a field for which you can't decide a name (e.g. `Distance` to `distance` created with `Distance matrix`)
It would be great to be able to do such action without having to overburden the model (see below).
**Additional context**
There are some alternatives, as discussed on this SO post : [https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler](https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler)
Yet, using the `Refactoring fields` algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler.
Currently, the best alternative is to add several steps to the process : create a temporary new field, copy the desired value in this field, then rename the temporary field.
| process | case sensitive field calculator using graphical modeler feature description at the moment field calculator does not seem to be case sensitive when creating a new field therefore it is not possible to simply rename a field for which you can t decide a name e g distance to distance created with distance matrix it would be great to be able to do such action without having to overburden the model see below additional context there are some alternatives as discussed on this so post yet using the refactoring fields algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler currently the best alternative is to add several steps to the process create a temporary new field copy the desired value in this field then rename the temporary field | 1 |
3,664 | 6,694,769,035 | IssuesEvent | 2017-10-10 04:19:28 | york-region-tpss/stp | https://api.github.com/repos/york-region-tpss/stp | opened | View Watering Assignment - Display the Watering Assignment | enhancement process workflow | Record the date and time in a new table.
Display the date and time associated with assignment number in the select list. | 1.0 | View Watering Assignment - Display the Watering Assignment - Record the date and time in a new table.
Display the date and time associated with assignment number in the select list. | process | view watering assignment display the watering assignment record the date and time in a new table display the date and time associated with assignment number in the select list | 1 |
127,705 | 27,105,237,104 | IssuesEvent | 2023-02-15 11:36:21 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Typo in "Your first Flutter app" step 6 | d: codelabs documentation | Issue for this codelab: https://codelabs.developers.google.com/codelabs/flutter-codelab-first#5
Very small typo that might, however, take your code out of sync with the example:
```
Next, add the Like button and connect it to toggleFavorites(). For a challenge, first try to do this by yourself, without looking at the code block below.
```
The function name reference here should be `toggleFavorite()`, not Favorites. | 1.0 | Typo in "Your first Flutter app" step 6 - Issue for this codelab: https://codelabs.developers.google.com/codelabs/flutter-codelab-first#5
Very small typo that might, however, take your code out of sync with the example:
```
Next, add the Like button and connect it to toggleFavorites(). For a challenge, first try to do this by yourself, without looking at the code block below.
```
The function name reference here should be `toggleFavorite()`, not Favorites. | non_process | typo in your first flutter app step issue for this codelab very small typo that might however take your code out of sync with the example next add the like button and connect it to togglefavorites for a challenge first try to do this by yourself without looking at the code block below the function name reference here should be togglefavorite not favorites | 0 |
659,547 | 21,932,515,066 | IssuesEvent | 2022-05-23 11:01:41 | PerfectFit-project/virtual-coach-issues | https://api.github.com/repos/PerfectFit-project/virtual-coach-issues | closed | Implement part of future-self dialog in botkit | spike high priority | Follow-up from #250. Botkit seems perfect for our type of dialogs. But let's investigate how good the fit is if we want to go into this path. | 1.0 | Implement part of future-self dialog in botkit - Follow-up from #250. Botkit seems perfect for our type of dialogs. But let's investigate how good the fit is if we want to go into this path. | non_process | implement part of future self dialog in botkit follow up from botkit seems perfect for our type of dialogs but let s investigate how good the fit is if we want to go into this path | 0 |
111,142 | 17,015,798,746 | IssuesEvent | 2021-07-02 11:51:49 | anyulled/react-playground | https://api.github.com/repos/anyulled/react-playground | opened | CVE-2021-29059 (High) detected in is-svg-3.0.0.tgz | security vulnerability | ## CVE-2021-29059 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: react-playground/package.json</p>
<p>Path to vulnerable library: react-playground/node_modules/is-svg</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- optimize-css-assets-webpack-plugin-5.0.4.tgz
- cssnano-4.1.10.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/anyulled/react-playground/commit/523de5a20017f19a3c0d97781ef28d338d041797">523de5a20017f19a3c0d97781ef28d338d041797</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was discovered in IS-SVG version 4.3.1 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string.
<p>Publish Date: 2021-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-29059 (High) detected in is-svg-3.0.0.tgz - ## CVE-2021-29059 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: react-playground/package.json</p>
<p>Path to vulnerable library: react-playground/node_modules/is-svg</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- optimize-css-assets-webpack-plugin-5.0.4.tgz
- cssnano-4.1.10.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/anyulled/react-playground/commit/523de5a20017f19a3c0d97781ef28d338d041797">523de5a20017f19a3c0d97781ef28d338d041797</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was discovered in IS-SVG version 4.3.1 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string.
<p>Publish Date: 2021-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in is svg tgz cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file react playground package json path to vulnerable library react playground node modules is svg dependency hierarchy react scripts tgz root library optimize css assets webpack plugin tgz cssnano tgz cssnano preset default tgz postcss svgo tgz x is svg tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was discovered in is svg version and below where a regular expression denial of service redos occurs if the application is provided and checks a crafted invalid svg string publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.