Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
212,693
16,473,470,212
IssuesEvent
2021-05-23 21:48:20
reclosedev/requests-cache
https://api.github.com/repos/reclosedev/requests-cache
closed
Run tests against multiple versions of requests library
tests
It would be worthwhile to run tests against multiple versions of `requests`, let's say the previous 4 (or more?) minor versions in addition to the latest version. Currently this would be `2.25.1` (latest), `2.24.0`, `2.23.0`, `2.22.0`, `2.21.0`. This should be relatively simple using a version matrix in GitHub Actions. Since this would add some time to CI jobs (even if run in parallel), maybe this should only be run for merges into `master` and skipped for other builds like pull requests.
1.0
Run tests against multiple versions of requests library - It would be worthwhile to run tests against multiple versions of `requests`, let's say the previous 4 (or more?) minor versions in addition to the latest version. Currently this would be `2.25.1` (latest), `2.24.0`, `2.23.0`, `2.22.0`, `2.21.0`. This should be relatively simple using a version matrix in GitHub Actions. Since this would add some time to CI jobs (even if run in parallel), maybe this should only be run for merges into `master` and skipped for other builds like pull requests.
non_process
run tests against multiple versions of requests library it would be worthwhile to run tests against multiple versions of requests let s say the previous or more minor versions in addition to the latest version currently this would be latest this should be relatively simple using a version matrix in github actions since this would add some time to ci jobs even if run in parallel maybe this should only be run for merges into master and skipped for other builds like pull requests
0
115,019
9,778,827,319
IssuesEvent
2019-06-07 13:06:16
ValveSoftware/halflife
https://api.github.com/repos/ValveSoftware/halflife
closed
[CS 1.6][REQ] Music in MOTD
Need Retest reviewed
Another anti-annoying-music request. Thank you for adding _speak_enable_ cvar, it works perfectly, but there's an another way to play high volume music on 1.6 server. With [this code](http://pastebin.com/Uhq2tvQm) in MOTD all players will listen for music if they have **volume 0**, **MP3Volume 0** and even after they leave your server. Song will play until it ends or till you quit cs. Windows 7 x86, build 5958.
1.0
[CS 1.6][REQ] Music in MOTD - Another anti-annoying-music request. Thank you for adding _speak_enable_ cvar, it works perfectly, but there's an another way to play high volume music on 1.6 server. With [this code](http://pastebin.com/Uhq2tvQm) in MOTD all players will listen for music if they have **volume 0**, **MP3Volume 0** and even after they leave your server. Song will play until it ends or till you quit cs. Windows 7 x86, build 5958.
non_process
music in motd another anti annoying music request thank you for adding speak enable cvar it works perfectly but there s an another way to play high volume music on server with in motd all players will listen for music if they have volume and even after they leave your server song will play until it ends or till you quit cs windows build
0
32,609
12,133,323,307
IssuesEvent
2020-04-23 08:50:45
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
JIT policy is not clear on usage
Pri1 assigned-to-author doc-enhancement security-center/svc triaged
When creating a policy that controls access to RDP port 3389 for the CIDR block of 10.21.254.0/24 I would expect that anyone making a JIT access request from within this range would be able to do so, e.g requesting connections from 10.21.254.85 to be allowed. This, however, is not the case, the only allowed connection from my experiments is when you request for the entire block to allowed all at once, 10.21.254.0/24. This to me doesn't really make a lot of sense, and I'm pretty sure this is not the intention of JIT or am i mistaken and the JIT policy has to be matched verbatim? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25f0758b-8274-d2b5-4702-2c7ca6b0b14a * Version Independent ID: d1215107-abf9-fefd-a63b-b20ec3f7f744 * Content: [Just-in-time virtual machine access in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-just-in-time#feedback) * Content Source: [articles/security-center/security-center-just-in-time.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-just-in-time.md) * Service: **security-center** * GitHub Login: @memildin * Microsoft Alias: **memildin**
True
JIT policy is not clear on usage - When creating a policy that controls access to RDP port 3389 for the CIDR block of 10.21.254.0/24 I would expect that anyone making a JIT access request from within this range would be able to do so, e.g requesting connections from 10.21.254.85 to be allowed. This, however, is not the case, the only allowed connection from my experiments is when you request for the entire block to allowed all at once, 10.21.254.0/24. This to me doesn't really make a lot of sense, and I'm pretty sure this is not the intention of JIT or am i mistaken and the JIT policy has to be matched verbatim? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25f0758b-8274-d2b5-4702-2c7ca6b0b14a * Version Independent ID: d1215107-abf9-fefd-a63b-b20ec3f7f744 * Content: [Just-in-time virtual machine access in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-just-in-time#feedback) * Content Source: [articles/security-center/security-center-just-in-time.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-just-in-time.md) * Service: **security-center** * GitHub Login: @memildin * Microsoft Alias: **memildin**
non_process
jit policy is not clear on usage when creating a policy that controls access to rdp port for the cidr block of i would expect that anyone making a jit access request from within this range would be able to do so e g requesting connections from to be allowed this however is not the case the only allowed connection from my experiments is when you request for the entire block to allowed all at once this to me doesn t really make a lot of sense and i m pretty sure this is not the intention of jit or am i mistaken and the jit policy has to be matched verbatim document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id fefd content content source service security center github login memildin microsoft alias memildin
0
56,183
11,525,125,113
IssuesEvent
2020-02-15 05:51:13
jsbxyyx/x-comment
https://api.github.com/repos/jsbxyyx/x-comment
opened
leetcode 0005 最长回文子串 | xxzkid | 逗的8笔
Gitalk leetcode-0005
https://uyiplus.com/2020/leetcode-0005/ 给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 1000。 示例 1:123输入: "babad"输出: "bab"注意: "aba" 也是一个有效答案。 示例 2:12输入: "cbbd"输出: "bb"
1.0
leetcode 0005 最长回文子串 | xxzkid | 逗的8笔 - https://uyiplus.com/2020/leetcode-0005/ 给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 1000。 示例 1:123输入: "babad"输出: "bab"注意: "aba" 也是一个有效答案。 示例 2:12输入: "cbbd"输出: "bb"
non_process
leetcode 最长回文子串 xxzkid 给定一个字符串 s,找到 s 中最长的回文子串。你可以假设 s 的最大长度为 。 示例 : quot babad quot 输出 quot bab quot 注意 quot aba quot 也是一个有效答案。 示例 : quot cbbd quot 输出 quot bb quot
0
184,261
14,284,063,288
IssuesEvent
2020-11-23 11:57:19
rtCamp/rtMedia
https://api.github.com/repos/rtCamp/rtMedia
closed
[Dashbaord] Hire Us tab - Correct the back ground color of the button
Low Tested UI
It should be `#007cba` color. ![Hire_Us_‹_test_—_WordPress](https://user-images.githubusercontent.com/15173772/69696206-fd928200-1104-11ea-904b-f710c3963668.png)
1.0
[Dashbaord] Hire Us tab - Correct the back ground color of the button - It should be `#007cba` color. ![Hire_Us_‹_test_—_WordPress](https://user-images.githubusercontent.com/15173772/69696206-fd928200-1104-11ea-904b-f710c3963668.png)
non_process
hire us tab correct the back ground color of the button it should be color
0
32,121
6,717,569,809
IssuesEvent
2017-10-14 23:04:28
k3dar/rdiff-backup-fs
https://api.github.com/repos/k3dar/rdiff-backup-fs
closed
Can't compile
auto-migrated Priority-Medium Type-Defect
_From @GoogleCodeExporter on July 5, 2015 0:23_ ``` Hi, I'm using current debian testing (squeeze). Here's what I get after ./configure while trying to make: gcc -c -Wall `pkg-config --cflags fuse` -D_GNU_SOURCE archfs.c /bin/sh: pkg-config: not found In file included from /usr/include/fuse/fuse.h:26, from /usr/include/fuse.h:9, from headers.h:21, from archfs.c:1: /usr/include/fuse/fuse_common.h:32:2: error: #error Please add -D_FILE_OFFSET_BITS=64 to your compile flags! make: *** [archfs.o] Error 1 Configure output: checking for gcc... gcc checking whether the C compiler works... yes checking for C compiler default output file name... a.out checking for suffix of executables... checking whether we are cross compiling... no checking for suffix of object files... o checking whether we are using the GNU C compiler... yes checking whether gcc accepts -g... yes checking for gcc option to accept ISO C89... none needed checking for fuse_main in -lfuse... yes checking how to run the C preprocessor... gcc -E checking for grep that handles long lines and -e... /bin/grep checking for egrep... /bin/grep -E checking for ANSI C header files... yes checking for sys/types.h... yes checking for sys/stat.h... yes checking for stdlib.h... yes checking for string.h... yes checking for memory.h... yes checking for strings.h... yes checking for inttypes.h... yes checking for stdint.h... yes checking for unistd.h... yes checking zlib.h usability... yes checking zlib.h presence... yes checking for zlib.h... yes checking stdio.h usability... yes checking stdio.h presence... yes checking for stdio.h... yes checking errno.h usability... yes checking errno.h presence... yes checking for errno.h... yes checking stdarg.h usability... yes checking stdarg.h presence... yes checking for stdarg.h... yes checking dirent.h usability... yes checking dirent.h presence... yes checking for dirent.h... yes checking sys/wait.h usability... yes checking sys/wait.h presence... yes checking for sys/wait.h... yes checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for get_current_dir_name... yes checking for getline... yes configure: creating ./config.status ``` Original issue reported on code.google.com by `snajdr.p...@gmail.com` on 29 Dec 2010 at 2:41 _Copied from original issue: rbrito/rdiff-backup-fs#1_
1.0
Can't compile - _From @GoogleCodeExporter on July 5, 2015 0:23_ ``` Hi, I'm using current debian testing (squeeze). Here's what I get after ./configure while trying to make: gcc -c -Wall `pkg-config --cflags fuse` -D_GNU_SOURCE archfs.c /bin/sh: pkg-config: not found In file included from /usr/include/fuse/fuse.h:26, from /usr/include/fuse.h:9, from headers.h:21, from archfs.c:1: /usr/include/fuse/fuse_common.h:32:2: error: #error Please add -D_FILE_OFFSET_BITS=64 to your compile flags! make: *** [archfs.o] Error 1 Configure output: checking for gcc... gcc checking whether the C compiler works... yes checking for C compiler default output file name... a.out checking for suffix of executables... checking whether we are cross compiling... no checking for suffix of object files... o checking whether we are using the GNU C compiler... yes checking whether gcc accepts -g... yes checking for gcc option to accept ISO C89... none needed checking for fuse_main in -lfuse... yes checking how to run the C preprocessor... gcc -E checking for grep that handles long lines and -e... /bin/grep checking for egrep... /bin/grep -E checking for ANSI C header files... yes checking for sys/types.h... yes checking for sys/stat.h... yes checking for stdlib.h... yes checking for string.h... yes checking for memory.h... yes checking for strings.h... yes checking for inttypes.h... yes checking for stdint.h... yes checking for unistd.h... yes checking zlib.h usability... yes checking zlib.h presence... yes checking for zlib.h... yes checking stdio.h usability... yes checking stdio.h presence... yes checking for stdio.h... yes checking errno.h usability... yes checking errno.h presence... yes checking for errno.h... yes checking stdarg.h usability... yes checking stdarg.h presence... yes checking for stdarg.h... yes checking dirent.h usability... yes checking dirent.h presence... yes checking for dirent.h... yes checking sys/wait.h usability... yes checking sys/wait.h presence... yes checking for sys/wait.h... yes checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for get_current_dir_name... yes checking for getline... yes configure: creating ./config.status ``` Original issue reported on code.google.com by `snajdr.p...@gmail.com` on 29 Dec 2010 at 2:41 _Copied from original issue: rbrito/rdiff-backup-fs#1_
non_process
can t compile from googlecodeexporter on july hi i m using current debian testing squeeze here s what i get after configure while trying to make gcc c wall pkg config cflags fuse d gnu source archfs c bin sh pkg config not found in file included from usr include fuse fuse h from usr include fuse h from headers h from archfs c usr include fuse fuse common h error error please add d file offset bits to your compile flags make error configure output checking for gcc gcc checking whether the c compiler works yes checking for c compiler default output file name a out checking for suffix of executables checking whether we are cross compiling no checking for suffix of object files o checking whether we are using the gnu c compiler yes checking whether gcc accepts g yes checking for gcc option to accept iso none needed checking for fuse main in lfuse yes checking how to run the c preprocessor gcc e checking for grep that handles long lines and e bin grep checking for egrep bin grep e checking for ansi c header files yes checking for sys types h yes checking for sys stat h yes checking for stdlib h yes checking for string h yes checking for memory h yes checking for strings h yes checking for inttypes h yes checking for stdint h yes checking for unistd h yes checking zlib h usability yes checking zlib h presence yes checking for zlib h yes checking stdio h usability yes checking stdio h presence yes checking for stdio h yes checking errno h usability yes checking errno h presence yes checking for errno h yes checking stdarg h usability yes checking stdarg h presence yes checking for stdarg h yes checking dirent h usability yes checking dirent h presence yes checking for dirent h yes checking sys wait h usability yes checking sys wait h presence yes checking for sys wait h yes checking pthread h usability yes checking pthread h presence yes checking for pthread h yes checking for get current dir name yes checking for getline yes configure creating config status original issue reported on code google com by snajdr p gmail com on dec at copied from original issue rbrito rdiff backup fs
0
65,461
12,597,829,650
IssuesEvent
2020-06-11 01:04:45
grpc/grpc-java
https://api.github.com/repos/grpc/grpc-java
opened
xds: investigate io.envoyproxy.controlplane:api
code health
Similar to #6016 which was not possible because the latest version of `io.envoyproxy.protoc-gen-validate:pgv-java-stub` in maven is not up to date as what we needed. `io.envoyproxy.controlplane:api` in maven is more up to date and even includes `com.google.cloudapi` that #7109 needs. However, we don't know how frequently it updates on a regular basis. It may still fall behind our need.
1.0
xds: investigate io.envoyproxy.controlplane:api - Similar to #6016 which was not possible because the latest version of `io.envoyproxy.protoc-gen-validate:pgv-java-stub` in maven is not up to date as what we needed. `io.envoyproxy.controlplane:api` in maven is more up to date and even includes `com.google.cloudapi` that #7109 needs. However, we don't know how frequently it updates on a regular basis. It may still fall behind our need.
non_process
xds investigate io envoyproxy controlplane api similar to which was not possible because the latest version of io envoyproxy protoc gen validate pgv java stub in maven is not up to date as what we needed io envoyproxy controlplane api in maven is more up to date and even includes com google cloudapi that needs however we don t know how frequently it updates on a regular basis it may still fall behind our need
0
87,857
10,981,936,497
IssuesEvent
2019-12-01 02:41:43
gopherjs/vecty
https://api.github.com/repos/gopherjs/vecty
closed
vecty.Render into non-body component
arch-design enhancement
Currently Render is restricted to a body component; i.e. Vecty must own the whole page effectively. We should allow rendering directly into an element that we own, e.g. `vecty.RenderInto("elem-id")` or something.
1.0
vecty.Render into non-body component - Currently Render is restricted to a body component; i.e. Vecty must own the whole page effectively. We should allow rendering directly into an element that we own, e.g. `vecty.RenderInto("elem-id")` or something.
non_process
vecty render into non body component currently render is restricted to a body component i e vecty must own the whole page effectively we should allow rendering directly into an element that we own e g vecty renderinto elem id or something
0
804,730
29,498,635,953
IssuesEvent
2023-06-02 19:21:31
asc-csa/Scisat-App
https://api.github.com/repos/asc-csa/Scisat-App
closed
Missing definition of error bars in altitude plot figure caption
science low priority
The figure caption should indicate the type of error bars. With further edits of the text: Graph [Plot] showing the gas concentration in parts per volume (ppv) over the selected altitude interval. The value represents the mean concentration over the selected latitudes, longitudes and dates. Error bars are 95% confidence intervals around the mean.
1.0
Missing definition of error bars in altitude plot figure caption - The figure caption should indicate the type of error bars. With further edits of the text: Graph [Plot] showing the gas concentration in parts per volume (ppv) over the selected altitude interval. The value represents the mean concentration over the selected latitudes, longitudes and dates. Error bars are 95% confidence intervals around the mean.
non_process
missing definition of error bars in altitude plot figure caption the figure caption should indicate the type of error bars with further edits of the text graph showing the gas concentration in parts per volume ppv over the selected altitude interval the value represents the mean concentration over the selected latitudes longitudes and dates error bars are confidence intervals around the mean
0
15,137
11,371,924,594
IssuesEvent
2020-01-28 00:01:46
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Unpin Microsoft.Bcl.AsyncInterfaces dependency
area-infrastructure
We have [one pinned dependency on dotnet/corefx](https://github.com/dotnet/aspnetcore/blob/2cc333ac3013218a3985ecd0809b0c541cbe1f91/eng/Version.Details.xml#L405). It means our dependency flow widget has a big red box we can't get rid of :( ![image](https://user-images.githubusercontent.com/7574/73203739-7f496080-40f2-11ea-8637-bbd25aa40c23.png) The dependency is `Microsoft.Bcl.AsyncInterfaces` (needed to build `IAsyncEnumerable` stuff on .NET Standard). It was pinned in https://github.com/dotnet/aspnetcore/pull/16716 by @BrennanConroy. @BrennanConroy do you know if we can unpin now?
1.0
Unpin Microsoft.Bcl.AsyncInterfaces dependency - We have [one pinned dependency on dotnet/corefx](https://github.com/dotnet/aspnetcore/blob/2cc333ac3013218a3985ecd0809b0c541cbe1f91/eng/Version.Details.xml#L405). It means our dependency flow widget has a big red box we can't get rid of :( ![image](https://user-images.githubusercontent.com/7574/73203739-7f496080-40f2-11ea-8637-bbd25aa40c23.png) The dependency is `Microsoft.Bcl.AsyncInterfaces` (needed to build `IAsyncEnumerable` stuff on .NET Standard). It was pinned in https://github.com/dotnet/aspnetcore/pull/16716 by @BrennanConroy. @BrennanConroy do you know if we can unpin now?
non_process
unpin microsoft bcl asyncinterfaces dependency we have it means our dependency flow widget has a big red box we can t get rid of the dependency is microsoft bcl asyncinterfaces needed to build iasyncenumerable stuff on net standard it was pinned in by brennanconroy brennanconroy do you know if we can unpin now
0
18,939
24,900,034,991
IssuesEvent
2022-10-28 19:48:05
MicrosoftDocs/windows-dev-docs
https://api.github.com/repos/MicrosoftDocs/windows-dev-docs
closed
Hiding windowsupdate-optionalupdates hides all windows update panel not only optional updates
uwp/prod processes-and-threading/tech Pri1
Hi, If I want to hide the optional updates part from Windows Update panel, it will not only hide optional updates but all Windows Update panel. The URI to use is the following one: windowsupdate-optionalupdates I try with CSP and catalog settings as below: hide:windowsupdate-optionalupdates --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 987ec16c-9456-93a4-177a-dbd563be7eb7 * Version Independent ID: f41f0344-f7f6-f092-a6bf-fc4184a9b460 * Content: [Launch the Windows Settings app - UWP applications](https://learn.microsoft.com/en-us/windows/uwp/launch-resume/launch-settings-app#ms-settings-uri-scheme-reference) * Content Source: [windows-apps-src/launch-resume/launch-settings-app.md](https://github.com/MicrosoftDocs/windows-dev-docs/blob/docs/windows-apps-src/launch-resume/launch-settings-app.md) * Product: **uwp** * Technology: **processes-and-threading** * GitHub Login: @alvinashcraft * Microsoft Alias: **aashcraft**
1.0
Hiding windowsupdate-optionalupdates hides all windows update panel not only optional updates - Hi, If I want to hide the optional updates part from Windows Update panel, it will not only hide optional updates but all Windows Update panel. The URI to use is the following one: windowsupdate-optionalupdates I try with CSP and catalog settings as below: hide:windowsupdate-optionalupdates --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 987ec16c-9456-93a4-177a-dbd563be7eb7 * Version Independent ID: f41f0344-f7f6-f092-a6bf-fc4184a9b460 * Content: [Launch the Windows Settings app - UWP applications](https://learn.microsoft.com/en-us/windows/uwp/launch-resume/launch-settings-app#ms-settings-uri-scheme-reference) * Content Source: [windows-apps-src/launch-resume/launch-settings-app.md](https://github.com/MicrosoftDocs/windows-dev-docs/blob/docs/windows-apps-src/launch-resume/launch-settings-app.md) * Product: **uwp** * Technology: **processes-and-threading** * GitHub Login: @alvinashcraft * Microsoft Alias: **aashcraft**
process
hiding windowsupdate optionalupdates hides all windows update panel not only optional updates hi if i want to hide the optional updates part from windows update panel it will not only hide optional updates but all windows update panel the uri to use is the following one windowsupdate optionalupdates i try with csp and catalog settings as below hide windowsupdate optionalupdates document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product uwp technology processes and threading github login alvinashcraft microsoft alias aashcraft
1
3,197
4,236,926,162
IssuesEvent
2016-07-05 20:03:47
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
Prepare System.Security.Cryptography.Native for OpenSSL 1.1 release
System.Security X-Plat
OpenSSL 1.1 has a new model of opaque structures with more accessor functions. Anything previously obtained by an accessor function will likely be fine, but direct field accesses won't work. For example, UpRefEvpPkey will need to change to EVP_PKEY_up_ref, but needs to do the current behavior on 1.0.x builds.
True
Prepare System.Security.Cryptography.Native for OpenSSL 1.1 release - OpenSSL 1.1 has a new model of opaque structures with more accessor functions. Anything previously obtained by an accessor function will likely be fine, but direct field accesses won't work. For example, UpRefEvpPkey will need to change to EVP_PKEY_up_ref, but needs to do the current behavior on 1.0.x builds.
non_process
prepare system security cryptography native for openssl release openssl has a new model of opaque structures with more accessor functions anything previously obtained by an accessor function will likely be fine but direct field accesses won t work for example uprefevppkey will need to change to evp pkey up ref but needs to do the current behavior on x builds
0
241,526
26,256,819,694
IssuesEvent
2023-01-06 02:00:44
benlazarine/cas-overlay
https://api.github.com/repos/benlazarine/cas-overlay
opened
WS-2019-0490 (High) detected in jcommander-1.48.jar
security vulnerability
## WS-2019-0490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jcommander-1.48.jar</b></p></summary> <p>A Java framework to parse command line options with annotations.</p> <p>Library home page: <a href="http://beust.com/">http://beust.com/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/beust/jcommander/1.48/jcommander-1.48.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - pac4j-saml-3.4.0.jar - xmlsectool-2.0.0.jar - :x: **jcommander-1.48.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Inclusion of Functionality from Untrusted Control Sphere vulnerability found in jcommander before 1.75. jcommander resolving dependencies over HTTP instead of HTTPS. <p>Publish Date: 2019-02-19 <p>URL: <a href=https://github.com/cbeust/jcommander/commit/3ae95595febbed9c13f367b6bda5c0be1c572c53>WS-2019-0490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-19</p> <p>Fix Resolution (com.beust:jcommander): 1.75</p> <p>Direct dependency fix Resolution (org.apereo.cas:cas-server-support-oauth-webflow): 6.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0490 (High) detected in jcommander-1.48.jar - ## WS-2019-0490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jcommander-1.48.jar</b></p></summary> <p>A Java framework to parse command line options with annotations.</p> <p>Library home page: <a href="http://beust.com/">http://beust.com/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/beust/jcommander/1.48/jcommander-1.48.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - pac4j-saml-3.4.0.jar - xmlsectool-2.0.0.jar - :x: **jcommander-1.48.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Inclusion of Functionality from Untrusted Control Sphere vulnerability found in jcommander before 1.75. jcommander resolving dependencies over HTTP instead of HTTPS. <p>Publish Date: 2019-02-19 <p>URL: <a href=https://github.com/cbeust/jcommander/commit/3ae95595febbed9c13f367b6bda5c0be1c572c53>WS-2019-0490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-19</p> <p>Fix Resolution (com.beust:jcommander): 1.75</p> <p>Direct dependency fix Resolution (org.apereo.cas:cas-server-support-oauth-webflow): 6.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in jcommander jar ws high severity vulnerability vulnerable library jcommander jar a java framework to parse command line options with annotations library home page a href path to dependency file pom xml path to vulnerable library root repository com beust jcommander jcommander jar dependency hierarchy cas server support oauth webflow jar root library saml jar xmlsectool jar x jcommander jar vulnerable library vulnerability details inclusion of functionality from untrusted control sphere vulnerability found in jcommander before jcommander resolving dependencies over http instead of https publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com beust jcommander direct dependency fix resolution org apereo cas cas server support oauth webflow step up your open source security game with mend
0
737,513
25,519,235,212
IssuesEvent
2022-11-28 18:58:25
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[opentitantool] Persistent sessions
Component:Software Component:Tooling Priority:P3 Type:Enhancement SW:opentitantool
# Problem OpenTitan tool allows for a variety of backend "Transports", some of which are drivers for hardware debuggers, others emulate OpenTitan or related chips. By their nature, some of those drivers will maintain state in the Rust struct that implements the Transport trait, others may have USB connections to hardware, which gets disturbed if the USB connection is closed and restablished. Rust programs which link directly with OpenTitan lib are able to keep the Transport object for a long as needed, making a sequence of method invocations to "stimulate" the device under test. However, users of the command line OpenTitan tool does not have that ability, since each command line invocation generally results in the creation of a Transport object, and the one "useful" method invocation, before the Transport is closed as the process terminates. # Proposal We will add a new optional persistent “OpenTitan session proxy” process (a daemon).  When used, all communication with the Transport backend will go through this proxy process. The proxy process will listen on a TCP socket and contain the "real" Transport implementation, maintaining emulator state or connection to USB hardware. The command line OpenTitan tool will use a "stub" Transport implementation, which forwards all requests via a TCP connection to the proxy. Design document: https://docs.google.com/document/d/18JIqdYcmJKDzANheNS9JI1rEre6-iXPZhjMjliAS0TE/edit# ## Tasks - [x] Make sure all error return values from Transport methods can be serialized (with serde) (#10964) - [x] Declare wire protocol by as a number of Rust structs/enums (#10994) - [x] Implement "stub" Transport (using serde JSON text via TCP) (#11023) - [x] Implement proxy server functionality (opentitansession binary) (#11054) - [x] Ability for opentitansession to run in background (as daemon) (#11173) This proposal has significant overlap with previous issue #10217, which proposed a JSON TCP protocol for communicating with an emulator running in a separate process. With this proposal, the interface to the emulator would instead be the Transport trait itself, that is, the top-level process of the emulator would either be opentitansession, or would be another binary that links with a library containing most of the functionality of opentitansession (the listening on TCP socket, and translation of protocol into calls to Transport trait.)
1.0
[opentitantool] Persistent sessions - # Problem OpenTitan tool allows for a variety of backend "Transports", some of which are drivers for hardware debuggers, others emulate OpenTitan or related chips. By their nature, some of those drivers will maintain state in the Rust struct that implements the Transport trait, others may have USB connections to hardware, which gets disturbed if the USB connection is closed and restablished. Rust programs which link directly with OpenTitan lib are able to keep the Transport object for a long as needed, making a sequence of method invocations to "stimulate" the device under test. However, users of the command line OpenTitan tool does not have that ability, since each command line invocation generally results in the creation of a Transport object, and the one "useful" method invocation, before the Transport is closed as the process terminates. # Proposal We will add a new optional persistent “OpenTitan session proxy” process (a daemon).  When used, all communication with the Transport backend will go through this proxy process. The proxy process will listen on a TCP socket and contain the "real" Transport implementation, maintaining emulator state or connection to USB hardware. The command line OpenTitan tool will use a "stub" Transport implementation, which forwards all requests via a TCP connection to the proxy. Design document: https://docs.google.com/document/d/18JIqdYcmJKDzANheNS9JI1rEre6-iXPZhjMjliAS0TE/edit# ## Tasks - [x] Make sure all error return values from Transport methods can be serialized (with serde) (#10964) - [x] Declare wire protocol by as a number of Rust structs/enums (#10994) - [x] Implement "stub" Transport (using serde JSON text via TCP) (#11023) - [x] Implement proxy server functionality (opentitansession binary) (#11054) - [x] Ability for opentitansession to run in background (as daemon) (#11173) This proposal has significant overlap with previous issue #10217, which proposed a JSON TCP protocol for communicating with an emulator running in a separate process. With this proposal, the interface to the emulator would instead be the Transport trait itself, that is, the top-level process of the emulator would either be opentitansession, or would be another binary that links with a library containing most of the functionality of opentitansession (the listening on TCP socket, and translation of protocol into calls to Transport trait.)
non_process
persistent sessions problem opentitan tool allows for a variety of backend transports some of which are drivers for hardware debuggers others emulate opentitan or related chips by their nature some of those drivers will maintain state in the rust struct that implements the transport trait others may have usb connections to hardware which gets disturbed if the usb connection is closed and restablished rust programs which link directly with opentitan lib are able to keep the transport object for a long as needed making a sequence of method invocations to stimulate the device under test however users of the command line opentitan tool does not have that ability since each command line invocation generally results in the creation of a transport object and the one useful method invocation before the transport is closed as the process terminates proposal we will add a new optional persistent “opentitan session proxy” process a daemon   when used all communication with the transport backend will go through this proxy process the proxy process will listen on a tcp socket and contain the real transport implementation maintaining emulator state or connection to usb hardware the command line opentitan tool will use a stub transport implementation which forwards all requests via a tcp connection to the proxy design document tasks make sure all error return values from transport methods can be serialized with serde declare wire protocol by as a number of rust structs enums implement stub transport using serde json text via tcp implement proxy server functionality opentitansession binary ability for opentitansession to run in background as daemon this proposal has significant overlap with previous issue which proposed a json tcp protocol for communicating with an emulator running in a separate process with this proposal the interface to the emulator would instead be the transport trait itself that is the top level process of the emulator would either be opentitansession or would be another binary that links with a library containing most of the functionality of opentitansession the listening on tcp socket and translation of protocol into calls to transport trait
0
17,168
11,756,107,318
IssuesEvent
2020-03-13 10:51:57
CIRCL/AIL-framework
https://api.github.com/repos/CIRCL/AIL-framework
opened
Add an option in the onion explorer to hide the result when there is no screenshot
enhancement usability
Add an option in the onion explorer to hide the result when there is no screenshot
True
Add an option in the onion explorer to hide the result when there is no screenshot - Add an option in the onion explorer to hide the result when there is no screenshot
non_process
add an option in the onion explorer to hide the result when there is no screenshot add an option in the onion explorer to hide the result when there is no screenshot
0
22,648
31,895,827,328
IssuesEvent
2023-09-18 01:31:56
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - formation
Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete
## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_formation Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation * Term label (English, not normative): Formation * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
1.0
Change term - formation - ## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_formation Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation * Term label (English, not normative): Formation * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
process
change term formation term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes formation term label english not normative formation organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the lithostratigraphic formation from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative notch peak formation house limestone fillmore formation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
1
10,720
13,523,058,801
IssuesEvent
2020-09-15 09:25:12
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
opened
Move IO repos to sgkit
multi-repo process + tools
We've [decided](https://discourse.pystatgen.org/t/20200910-developer-call-notes/96) to move sgkit-{bgen,plink,vcf} to the main sgkit repo. This is an umbrella issue to track tasks. See also discussion in #65 - [ ] Add Windows CI #247 - [ ] Move sgkit-bgen to main sgkit repo #256 - [ ] Move sgkit-plink to main sgkit repo #257 - [ ] Move sgkit-vcf to main sgkit repo #258
1.0
Move IO repos to sgkit - We've [decided](https://discourse.pystatgen.org/t/20200910-developer-call-notes/96) to move sgkit-{bgen,plink,vcf} to the main sgkit repo. This is an umbrella issue to track tasks. See also discussion in #65 - [ ] Add Windows CI #247 - [ ] Move sgkit-bgen to main sgkit repo #256 - [ ] Move sgkit-plink to main sgkit repo #257 - [ ] Move sgkit-vcf to main sgkit repo #258
process
move io repos to sgkit we ve to move sgkit bgen plink vcf to the main sgkit repo this is an umbrella issue to track tasks see also discussion in add windows ci move sgkit bgen to main sgkit repo move sgkit plink to main sgkit repo move sgkit vcf to main sgkit repo
1
22,486
31,395,279,836
IssuesEvent
2023-08-26 21:22:57
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added)
suggested title in process
Please add as much of the following info as you can: Title: Joan is Awful Type (film/tv show): TV show - drama Film or show in which it appears: Black Mirror Is the parent film/show streaming anywhere? Yes - Netflix About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful" Actual footage of the film/show can be seen (yes/no)? Yes Timestamps (interspersed with people watching it): - 12:50 - 13:39 - 14:35 - 14:47 - 15:00 - 19:22 - 20:00 - 20:50 - 21:59 - 22:17 - 25:20 - 25:40 - 30:00 - 30:40 - 35:50 - 36:20 Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett Network: Streamberry (Streamberry CEO: Mona Javadi) Number of seasons: 1 Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing. ![Joan Thumbnail](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/16b534d8-b935-4595-be0f-f027ec715b3f) ![Joan Title](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/436ef050-2006-4cee-85b4-d21b5e903ab5) ![Joan1](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/bfa18db7-be73-4b0a-8006-31e6a7fd961d) ![Joan2](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/47a33c95-9c4d-421c-bf2c-41e1be55068f) ![Joan3](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7b5dfe5f-9749-45b5-992e-cd26eed91388) ![Joan4](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/c1e4c125-bf0e-4c60-a662-681eb93427ec) ![Joan5](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7370a8d2-6bb2-4314-81b2-56e6a52cc08e) ![Joan6](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0cfb1b8f-0151-4f2c-b0da-c145a3acad22) ![Joan7](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7d659805-d69b-4f69-b3e2-4905d4729fb6) ![Joan9](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2a08ea44-3506-4432-a0e1-dc9d374d0154) ![Joan11](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/941e9cc9-4596-46bc-83dc-e62516fae17c) ![Joan12](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0786abd0-8f1d-4b31-a2f2-0ed718aa5497) ![Joan13](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f2cd77c7-87b1-49c8-bd34-89dd4ee649c1) ![Joan14](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/03d45d0d-29d3-4f0c-bb42-79e8cebb0f92) ![Joan16](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/d00bc43a-203a-4acb-85be-212f687f98ef) ![Joan17](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2538cb51-63c4-4134-ae1c-757a9163b1c2) ![Joan19](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/e7ce6be7-e4fe-4570-9ea4-866444551e60) ![Joan20](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/65003790-cfb6-4c7f-9220-c4cdb9014e5f) ![Joan21](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/73551597-0385-460c-a3ff-17f470fd9ef4) ![Joan22](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/8f60e7e8-82dd-480f-9bd2-9d0164089135) ![Joan23](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/5bc36f97-0e3c-41b2-a786-220b42a92d80) ![Joan24](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/b7d56eda-35a0-47cd-864b-582c1e571b03) ![Joan25](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/fa3aa4c0-b381-402c-b509-8065ce99ba3a) ![Joan26](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/1a6a9c61-bf9d-48ab-8a57-c19e04b31ae2) ![Joan27](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/a7fd989c-4913-49c9-909b-2eaf2813be72) ![Joan28](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f88e8775-cb7b-4875-880c-c3d06c07d82e)
1.0
Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added) - Please add as much of the following info as you can: Title: Joan is Awful Type (film/tv show): TV show - drama Film or show in which it appears: Black Mirror Is the parent film/show streaming anywhere? Yes - Netflix About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful" Actual footage of the film/show can be seen (yes/no)? Yes Timestamps (interspersed with people watching it): - 12:50 - 13:39 - 14:35 - 14:47 - 15:00 - 19:22 - 20:00 - 20:50 - 21:59 - 22:17 - 25:20 - 25:40 - 30:00 - 30:40 - 35:50 - 36:20 Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett Network: Streamberry (Streamberry CEO: Mona Javadi) Number of seasons: 1 Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing. ![Joan Thumbnail](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/16b534d8-b935-4595-be0f-f027ec715b3f) ![Joan Title](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/436ef050-2006-4cee-85b4-d21b5e903ab5) ![Joan1](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/bfa18db7-be73-4b0a-8006-31e6a7fd961d) ![Joan2](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/47a33c95-9c4d-421c-bf2c-41e1be55068f) ![Joan3](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7b5dfe5f-9749-45b5-992e-cd26eed91388) ![Joan4](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/c1e4c125-bf0e-4c60-a662-681eb93427ec) ![Joan5](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7370a8d2-6bb2-4314-81b2-56e6a52cc08e) ![Joan6](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0cfb1b8f-0151-4f2c-b0da-c145a3acad22) ![Joan7](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7d659805-d69b-4f69-b3e2-4905d4729fb6) ![Joan9](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2a08ea44-3506-4432-a0e1-dc9d374d0154) ![Joan11](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/941e9cc9-4596-46bc-83dc-e62516fae17c) ![Joan12](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0786abd0-8f1d-4b31-a2f2-0ed718aa5497) ![Joan13](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f2cd77c7-87b1-49c8-bd34-89dd4ee649c1) ![Joan14](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/03d45d0d-29d3-4f0c-bb42-79e8cebb0f92) ![Joan16](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/d00bc43a-203a-4acb-85be-212f687f98ef) ![Joan17](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2538cb51-63c4-4134-ae1c-757a9163b1c2) ![Joan19](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/e7ce6be7-e4fe-4570-9ea4-866444551e60) ![Joan20](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/65003790-cfb6-4c7f-9220-c4cdb9014e5f) ![Joan21](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/73551597-0385-460c-a3ff-17f470fd9ef4) ![Joan22](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/8f60e7e8-82dd-480f-9bd2-9d0164089135) ![Joan23](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/5bc36f97-0e3c-41b2-a786-220b42a92d80) ![Joan24](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/b7d56eda-35a0-47cd-864b-582c1e571b03) ![Joan25](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/fa3aa4c0-b381-402c-b509-8065ce99ba3a) ![Joan26](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/1a6a9c61-bf9d-48ab-8a57-c19e04b31ae2) ![Joan27](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/a7fd989c-4913-49c9-909b-2eaf2813be72) ![Joan28](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f88e8775-cb7b-4875-880c-c3d06c07d82e)
process
add joan is awful from black mirror screenshots thumbnail and title card added please add as much of the following info as you can title joan is awful type film tv show tv show drama film or show in which it appears black mirror is the parent film show streaming anywhere yes netflix about when in the parent film show does it appear episode joan is awful actual footage of the film show can be seen yes no yes timestamps interspersed with people watching it starring salma hayek himesh patel jaboukie young white camirin farmer ben marnes cate blanchett network streamberry streamberry ceo mona javadi number of seasons note this episode of black mirror is crazy meta there s a version of joan is awful within the show of joan is awful and the entirity of the episode is revealed to be the real joan is awful show i m just putting the primary joan is awful show here in order to not be too confusing
1
12,692
15,058,119,550
IssuesEvent
2021-02-03 22:51:03
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Standard Fields - Edge Cases and Oddities
bug team:data processing
### Describe the bug Reviewing the schemas revealed some oddities and edge cases for the Standard Fields that Panther normalizes across all log records. ### Expected behavior The following are summaries of the anomalous behavior from the `log_processor` testdata. Refer to the **Additional Context** section for more details. #### Anomalous Behavior - GitLab - Production Tests<br />The `location` field (`"http://34.222.254.254/users/sign_in"`) shows an IP Address. <br />Shouldn't `p_any_ip_addresses` include `34.222.254.254`?<br /><br /> - GSuite - Data clearly shows domain names: <br />i.e. `www.google.com/accounts/OAuthLogin` and `77185425430.apps.googleusercontent.com` <br />However the results do not include any `p_any_domain_names`<br /><br /> - AWS - These values should be included in their respective `p_` fields but are not included: - `p_any_aws_arns` - `arn:aws:wafv2:us-west-2:123456789EXAMPLE:regional/webacl/panther_web_dev/f96fe2f4-a2f4-4c06-a4bc-afd495bd675f` - `p_any_domain_names` - `web-123456789.us-west-2.elb.amazonaws.com` - `p_any_aws_account_ids` - `123456789EXAMPLE` and `123456789EXAMPLE_MANAGED` #### Others - Consider `ip-172-31-14-137.us-west-2.compute.internal` <br />While this is explicitly a domain name, it also explicitly denotes an IP address. <br />Would we want to extract `172.31.14.137` as an IP address?<br /><br /> - From the Sophos logs: - There is a file path value: `"/Users/jsmith/Dropbox/.dropbox.cache/new_files/eec8381bba1914aa2207f4a53c0fad1e"`. <br />Based on the context of the test data, this appears to be a Win x32 system. <br />However this file path is not a valid Windows file path (would be valid if `/` were `\`) - In one of the tests, there is a file path explicitly stated in the `name`. <br />Since the file path input key is not present, the results do not show any file paths. <br />Would this make a case to add support for file paths in the standardized `p_` fields? - AWS - VPC DNS Tests - `p_any_domain_names` - Why is `xn--fa-hia.com` not included in this `p_` field? - What's up with the `"faß.com"` included in this `p_` field? - Why is `foo@bar.com` included in this `p_` field? - Shouldn't `foo@bar.com` be included in the `p_any_emails` field? - CloudTrail (base and insight) Tests - While not particularly useful, shouldn't `p_any_domain_names` match on the following: - `kms.amazonaws.com` - `ssm.amazonaws.com` ### Additional context See this [Pull Request Review](https://github.com/panther-labs/panther-analysis/pull/187#pullrequestreview-581874346) for full context.
1.0
Standard Fields - Edge Cases and Oddities - ### Describe the bug Reviewing the schemas revealed some oddities and edge cases for the Standard Fields that Panther normalizes across all log records. ### Expected behavior The following are summaries of the anomalous behavior from the `log_processor` testdata. Refer to the **Additional Context** section for more details. #### Anomalous Behavior - GitLab - Production Tests<br />The `location` field (`"http://34.222.254.254/users/sign_in"`) shows an IP Address. <br />Shouldn't `p_any_ip_addresses` include `34.222.254.254`?<br /><br /> - GSuite - Data clearly shows domain names: <br />i.e. `www.google.com/accounts/OAuthLogin` and `77185425430.apps.googleusercontent.com` <br />However the results do not include any `p_any_domain_names`<br /><br /> - AWS - These values should be included in their respective `p_` fields but are not included: - `p_any_aws_arns` - `arn:aws:wafv2:us-west-2:123456789EXAMPLE:regional/webacl/panther_web_dev/f96fe2f4-a2f4-4c06-a4bc-afd495bd675f` - `p_any_domain_names` - `web-123456789.us-west-2.elb.amazonaws.com` - `p_any_aws_account_ids` - `123456789EXAMPLE` and `123456789EXAMPLE_MANAGED` #### Others - Consider `ip-172-31-14-137.us-west-2.compute.internal` <br />While this is explicitly a domain name, it also explicitly denotes an IP address. <br />Would we want to extract `172.31.14.137` as an IP address?<br /><br /> - From the Sophos logs: - There is a file path value: `"/Users/jsmith/Dropbox/.dropbox.cache/new_files/eec8381bba1914aa2207f4a53c0fad1e"`. <br />Based on the context of the test data, this appears to be a Win x32 system. <br />However this file path is not a valid Windows file path (would be valid if `/` were `\`) - In one of the tests, there is a file path explicitly stated in the `name`. <br />Since the file path input key is not present, the results do not show any file paths. <br />Would this make a case to add support for file paths in the standardized `p_` fields? - AWS - VPC DNS Tests - `p_any_domain_names` - Why is `xn--fa-hia.com` not included in this `p_` field? - What's up with the `"faß.com"` included in this `p_` field? - Why is `foo@bar.com` included in this `p_` field? - Shouldn't `foo@bar.com` be included in the `p_any_emails` field? - CloudTrail (base and insight) Tests - While not particularly useful, shouldn't `p_any_domain_names` match on the following: - `kms.amazonaws.com` - `ssm.amazonaws.com` ### Additional context See this [Pull Request Review](https://github.com/panther-labs/panther-analysis/pull/187#pullrequestreview-581874346) for full context.
process
standard fields edge cases and oddities describe the bug reviewing the schemas revealed some oddities and edge cases for the standard fields that panther normalizes across all log records expected behavior the following are summaries of the anomalous behavior from the log processor testdata refer to the additional context section for more details anomalous behavior gitlab production tests the location field shows an ip address shouldn t p any ip addresses include gsuite data clearly shows domain names i e and apps googleusercontent com however the results do not include any p any domain names aws these values should be included in their respective p fields but are not included p any aws arns arn aws us west regional webacl panther web dev p any domain names web us west elb amazonaws com p any aws account ids and managed others consider ip us west compute internal while this is explicitly a domain name it also explicitly denotes an ip address would we want to extract as an ip address from the sophos logs there is a file path value users jsmith dropbox dropbox cache new files based on the context of the test data this appears to be a win system however this file path is not a valid windows file path would be valid if were in one of the tests there is a file path explicitly stated in the name since the file path input key is not present the results do not show any file paths would this make a case to add support for file paths in the standardized p fields aws vpc dns tests p any domain names why is xn fa hia com not included in this p field what s up with the faß com included in this p field why is foo bar com included in this p field shouldn t foo bar com be included in the p any emails field cloudtrail base and insight tests while not particularly useful shouldn t p any domain names match on the following kms amazonaws com ssm amazonaws com additional context see this for full context
1
2,836
5,791,834,513
IssuesEvent
2017-05-02 07:36:55
gaocegege/maintainer
https://api.github.com/repos/gaocegege/maintainer
opened
Transfer from builder pattern to multi-stage docker build
priority/P3 process/not claimed type/enhancement
ref http://blog.alexellis.io/mutli-stage-docker-builds/ Blocked until the feature is released. ref https://github.com/moby/moby/pull/31257 and https://github.com/moby/moby/pull/32063
1.0
Transfer from builder pattern to multi-stage docker build - ref http://blog.alexellis.io/mutli-stage-docker-builds/ Blocked until the feature is released. ref https://github.com/moby/moby/pull/31257 and https://github.com/moby/moby/pull/32063
process
transfer from builder pattern to multi stage docker build ref blocked until the feature is released ref and
1
120,708
25,850,910,844
IssuesEvent
2022-12-13 10:20:03
Clueless-Community/seamless-ui
https://api.github.com/repos/Clueless-Community/seamless-ui
closed
Improve Footer Dark 4
MEDIUM codepeak 22 issue:3
Need to improve this component on `Footer/src/footer-dark-04.html` ## Improvements required : The Button, Spacing, social media icons are not aligned as per figma file ## Any reference image? ![image](https://user-images.githubusercontent.com/98400348/207259112-51c960b8-6154-4749-b2e4-001fd62d53ef.png)
1.0
Improve Footer Dark 4 - Need to improve this component on `Footer/src/footer-dark-04.html` ## Improvements required : The Button, Spacing, social media icons are not aligned as per figma file ## Any reference image? ![image](https://user-images.githubusercontent.com/98400348/207259112-51c960b8-6154-4749-b2e4-001fd62d53ef.png)
non_process
improve footer dark need to improve this component on footer src footer dark html improvements required the button spacing social media icons are not aligned as per figma file any reference image
0
21,915
30,443,576,963
IssuesEvent
2023-07-15 11:29:14
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
skypilot-nightly 1.0.0.dev20230715 has 2 GuardDog issues
guarddog exec-base64 silent-process-execution
https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230715", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230715/sky/cloud_stores.py:110", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230715/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp79q_y85t/skypilot-nightly" } }```
1.0
skypilot-nightly 1.0.0.dev20230715 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230715", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230715/sky/cloud_stores.py:110", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230715/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp79q_y85t/skypilot-nightly" } }```
process
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp skypilot nightly
1
760,363
26,638,008,781
IssuesEvent
2023-01-25 00:17:22
ArjunSharda/Passeo
https://api.github.com/repos/ArjunSharda/Passeo
closed
Add more security to Passeo quickgenerate command
enhancement High Priority
This update will add more security to Passeo's quickgenerate's command.
1.0
Add more security to Passeo quickgenerate command - This update will add more security to Passeo's quickgenerate's command.
non_process
add more security to passeo quickgenerate command this update will add more security to passeo s quickgenerate s command
0
53,500
3,040,705,628
IssuesEvent
2015-08-07 16:52:11
scamille/simc_issue_test3
https://api.github.com/repos/scamille/simc_issue_test3
closed
Non-zero melee stat scale factors for Warlocks
bug imported Priority-Medium
_From [natehieter@gmail.com](https://code.google.com/u/natehieter@gmail.com/) on January 09, 2009 22:21:39_ When many Warlocks are added to the sim, some end up with apparent scaling against stats like expertise. Most likely, this is due to a spike in one or more iterations due to a bug. _Original issue: http://code.google.com/p/simulationcraft/issues/detail?id=18_
1.0
Non-zero melee stat scale factors for Warlocks - _From [natehieter@gmail.com](https://code.google.com/u/natehieter@gmail.com/) on January 09, 2009 22:21:39_ When many Warlocks are added to the sim, some end up with apparent scaling against stats like expertise. Most likely, this is due to a spike in one or more iterations due to a bug. _Original issue: http://code.google.com/p/simulationcraft/issues/detail?id=18_
non_process
non zero melee stat scale factors for warlocks from on january when many warlocks are added to the sim some end up with apparent scaling against stats like expertise most likely this is due to a spike in one or more iterations due to a bug original issue
0
9,616
12,553,266,190
IssuesEvent
2020-06-06 21:16:53
peopledoc/procrastinate
https://api.github.com/repos/peopledoc/procrastinate
closed
Auto-deploy to PyPI is still broken
Good for: newcomers Type: Bug Type: Process
It looks like our .travis.yml syntax is still broken, due to a mix between matrix and jobs.
1.0
Auto-deploy to PyPI is still broken - It looks like our .travis.yml syntax is still broken, due to a mix between matrix and jobs.
process
auto deploy to pypi is still broken it looks like our travis yml syntax is still broken due to a mix between matrix and jobs
1
1,565
10,343,139,407
IssuesEvent
2019-09-04 08:18:25
DimensionDev/Maskbook
https://api.github.com/repos/DimensionDev/Maskbook
closed
Just script: use of 'shell: true'
Dev: Local Automation Severity: Low Speedy Close Requested Type: Bug wontfix
> without this things won't work but it said this option is dangerous. -- @Artoria2e5
1.0
Just script: use of 'shell: true' - > without this things won't work but it said this option is dangerous. -- @Artoria2e5
non_process
just script use of shell true without this things won t work but it said this option is dangerous
0
147,264
11,780,316,280
IssuesEvent
2020-03-16 19:48:24
theophilusx/ssh2-sftp-client
https://api.github.com/repos/theophilusx/ssh2-sftp-client
closed
Reconnect on connection closed
bug test
We are building an upload tool that must be resilient to connection loss. Right now when the connection is lost while an append or put operation is in progress, we get a "close" event, but the append or put operations don't resolve or throw. Currently I'm wrapping our upload as such: ``` try { const success = await new Promise<boolean>(resolv => { const listener = (s: boolean) => { this.client.removeListener("close", listener); resolv(s); }; this.client.on("close", () => listener(false)); // If the connection is lost, neither of these ever resolve. if (tryCont) { this.client.append(stream, destFile).then(() => listener(true)); } else { this.client.put(stream, destFile).then(() => listener(true)); } }); if (!success) throw new Error("Connection lost"); } catch (e) { console.log(e); throw e; } ``` This works, but feels like its leaking in the append and put calls since they are not resolving. Along the same lines, calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of "no more sessons" errors when a connection is finally made. Its like each failed connect call finally makes a connection even when it reports failure when its initially called. Is there an intended approach for doing this kind of thing? Thanks!
1.0
Reconnect on connection closed - We are building an upload tool that must be resilient to connection loss. Right now when the connection is lost while an append or put operation is in progress, we get a "close" event, but the append or put operations don't resolve or throw. Currently I'm wrapping our upload as such: ``` try { const success = await new Promise<boolean>(resolv => { const listener = (s: boolean) => { this.client.removeListener("close", listener); resolv(s); }; this.client.on("close", () => listener(false)); // If the connection is lost, neither of these ever resolve. if (tryCont) { this.client.append(stream, destFile).then(() => listener(true)); } else { this.client.put(stream, destFile).then(() => listener(true)); } }); if (!success) throw new Error("Connection lost"); } catch (e) { console.log(e); throw e; } ``` This works, but feels like its leaking in the append and put calls since they are not resolving. Along the same lines, calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of "no more sessons" errors when a connection is finally made. Its like each failed connect call finally makes a connection even when it reports failure when its initially called. Is there an intended approach for doing this kind of thing? Thanks!
non_process
reconnect on connection closed we are building an upload tool that must be resilient to connection loss right now when the connection is lost while an append or put operation is in progress we get a close event but the append or put operations don t resolve or throw currently i m wrapping our upload as such try const success await new promise resolv const listener s boolean this client removelistener close listener resolv s this client on close listener false if the connection is lost neither of these ever resolve if trycont this client append stream destfile then listener true else this client put stream destfile then listener true if success throw new error connection lost catch e console log e throw e this works but feels like its leaking in the append and put calls since they are not resolving along the same lines calling connect multiple times to test for a connection results in the sftp server erroring out with a long list of no more sessons errors when a connection is finally made its like each failed connect call finally makes a connection even when it reports failure when its initially called is there an intended approach for doing this kind of thing thanks
0
20,188
26,753,344,520
IssuesEvent
2023-01-30 21:33:43
darkside-princeton/sipm-analysis
https://api.github.com/repos/darkside-princeton/sipm-analysis
closed
Rename SiPM and Dataset classes
pre-processing
Rename class SiPM to WaveformAnalyzer and class Dataset to WaveformDataset. Modify all files accordingly.
1.0
Rename SiPM and Dataset classes - Rename class SiPM to WaveformAnalyzer and class Dataset to WaveformDataset. Modify all files accordingly.
process
rename sipm and dataset classes rename class sipm to waveformanalyzer and class dataset to waveformdataset modify all files accordingly
1
35,527
4,995,422,933
IssuesEvent
2016-12-09 10:05:51
halestudio/hale
https://api.github.com/repos/halestudio/hale
closed
Arc interpolation takes very long or is not terminating
bug io to be tested
I have issues with the arc interpolation taking seemingly forever. I was not able to complete loading a project because loading the data did not finish in a sensible amount of time. refs #181 Internal example project `aaa-au-flurstuecke`. Please create a test case from the respective geometry causing this problem.
1.0
Arc interpolation takes very long or is not terminating - I have issues with the arc interpolation taking seemingly forever. I was not able to complete loading a project because loading the data did not finish in a sensible amount of time. refs #181 Internal example project `aaa-au-flurstuecke`. Please create a test case from the respective geometry causing this problem.
non_process
arc interpolation takes very long or is not terminating i have issues with the arc interpolation taking seemingly forever i was not able to complete loading a project because loading the data did not finish in a sensible amount of time refs internal example project aaa au flurstuecke please create a test case from the respective geometry causing this problem
0
162,379
20,188,185,937
IssuesEvent
2022-02-11 01:16:15
Nivaskumark/kernel_v4.19.72_old
https://api.github.com/repos/Nivaskumark/kernel_v4.19.72_old
opened
CVE-2022-0492 (High) detected in linuxlinux-4.19.83
security vulnerability
## CVE-2022-0492 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The cgroup release_agent is called with call_usermodehelper. The function call_usermodehelper starts the release_agent with a full set fo capabilities. Therefore require capabilities when setting the release_agaent. Before version v5.17-rc3 <p>Publish Date: 2022-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0492>CVE-2022-0492</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-0492">https://security-tracker.debian.org/tracker/CVE-2022-0492</a></p> <p>Release Date: 2022-02-04</p> <p>Fix Resolution: v5.17-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0492 (High) detected in linuxlinux-4.19.83 - ## CVE-2022-0492 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/cgroup/cgroup-v1.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The cgroup release_agent is called with call_usermodehelper. The function call_usermodehelper starts the release_agent with a full set fo capabilities. Therefore require capabilities when setting the release_agaent. Before version v5.17-rc3 <p>Publish Date: 2022-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0492>CVE-2022-0492</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2022-0492">https://security-tracker.debian.org/tracker/CVE-2022-0492</a></p> <p>Release Date: 2022-02-04</p> <p>Fix Resolution: v5.17-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files kernel cgroup cgroup c kernel cgroup cgroup c vulnerability details the cgroup release agent is called with call usermodehelper the function call usermodehelper starts the release agent with a full set fo capabilities therefore require capabilities when setting the release agaent before version publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
763,027
26,741,230,452
IssuesEvent
2023-01-30 13:08:15
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Provide custom host verifier implementation for pgjdbc
kind/bug priority/medium area/ecosystem
Jira Link: [DB-3807](https://yugabyte.atlassian.net/browse/DB-3807) ### Description In a YB Managed setup where a client is in a VPC peered with the YBDB cluster's VPC, the ssl connection with mode `verify-full` does not work. Provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by `yb_servers()` function.
1.0
[YSQL] Provide custom host verifier implementation for pgjdbc - Jira Link: [DB-3807](https://yugabyte.atlassian.net/browse/DB-3807) ### Description In a YB Managed setup where a client is in a VPC peered with the YBDB cluster's VPC, the ssl connection with mode `verify-full` does not work. Provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by `yb_servers()` function.
non_process
provide custom host verifier implementation for pgjdbc jira link description in a yb managed setup where a client is in a vpc peered with the ybdb cluster s vpc the ssl connection with mode verify full does not work provide a custom hostname verifier which can verify that the host address of the connection matches with one in the list returned by yb servers function
0
236,013
25,971,399,283
IssuesEvent
2022-12-19 11:31:32
nk7598/linux-4.19.72
https://api.github.com/repos/nk7598/linux-4.19.72
closed
CVE-2022-1998 (High) detected in linuxlinux-4.19.269 - autoclosed
security vulnerability
## CVE-2022-1998 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nk7598/linux-4.19.72/commit/8d6de636016872da224f31e7d9d0fe96d373b46c">8d6de636016872da224f31e7d9d0fe96d373b46c</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system. <p>Publish Date: 2022-06-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p> <p>Release Date: 2022-06-09</p> <p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1998 (High) detected in linuxlinux-4.19.269 - autoclosed - ## CVE-2022-1998 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.269</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nk7598/linux-4.19.72/commit/8d6de636016872da224f31e7d9d0fe96d373b46c">8d6de636016872da224f31e7d9d0fe96d373b46c</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system. <p>Publish Date: 2022-06-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p> <p>Release Date: 2022-06-09</p> <p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href vulnerable source files fs notify fanotify fanotify user c fs notify fanotify fanotify user c vulnerability details a use after free in the linux kernel file system notify functionality was found in the way user triggers copy info records to user call to fail in copy event to user a local user could use this flaw to crash the system or potentially escalate their privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
18,258
24,341,318,210
IssuesEvent
2022-10-01 18:47:49
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
closed
Argument list too long when connecting to a database with a large schema
bug/0-needs-info kind/bug process/candidate topic: client level/intermediate priority/high
## Bug description The query engine cannot spawn the query engine child process when the schema is too large. The error is coming from that line https://github.com/RobertCraigie/prisma-client-py/blob/main/src/prisma/generator/templates/engine/query.py.jinja#L106. When the schema is too large, it's content will be truncated by the OS and the engine will fail to start. The following patch will circumvent the issue by passing the path to the schema instead of it's content: ``` env = os.environ.copy() env.update( # PRISMA_DML=self.dml, RUST_LOG='error', RUST_LOG_FORMAT='json', PRISMA_CLIENT_ENGINE_TYPE='binary', ) if os.environ.get('PRISMA_DML_PATH'): env.update(PRISMA_DML_PATH=os.environ.get('PRISMA_DML_PATH')) else: env.update(PRISMA_DML=self.dml) ``` ## How to reproduce Steps to reproduce the behavior: 1. Generate a prisma.schema file with a large number of tables and fields 2. Generate the prisma client 3. Connect to the database 4. See error: ``` Traceback (most recent call last): File "...", line 239, in <module> app = CashflowLoader(args.process_date) File "...", line 35, in __init__ self.db.connect() File ".../venv/lib/python3.9/site-packages/prisma/client.py", line 3657, in connect self.__engine.connect( File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 110, in connect self.spawn(file, timeout=timeout, datasources=datasources) File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 152, in spawn self.process = subprocess.Popen( File ".../lib/python3.9/subprocess.py", line 951, in __init__ self._execute_child(args, executable, preexec_fn, close_fds, File ".../lib/python3.9/subprocess.py", line 1821, in _execute_child raise child_exception_type(errno_num, err_msg, err_filename) OSError: [Errno 7] Argument list too long: '/tmp/prisma/binaries/engines/efdf9b1183dddfd4258cd181a72125755215ab7b/prisma-query-engine-debian-openssl-1.1.x' ``` ## Expected behavior No error should be reported and the connexion should be established with the database. ## Prisma information Create a large schema file No specific queries required ## Environment & setup - OS: Windows 11, Ubuntu 20.10 - Database: MariaDB - Python version: 3.9.7 - Prisma version: ``` prisma : 3.13.0 prisma client python : 0.6.6 platform : debian-openssl-1.1.x engines : efdf9b1183dddfd4258cd181a72125755215ab7b install path : /mnt/c/Users/andrew/OneDrive/Git_Repos/e22/Benzaiten_repos/benzaiten-processes/venv/lib/python3.9/site-packages/prisma installed extras : [] ```
1.0
Argument list too long when connecting to a database with a large schema - ## Bug description The query engine cannot spawn the query engine child process when the schema is too large. The error is coming from that line https://github.com/RobertCraigie/prisma-client-py/blob/main/src/prisma/generator/templates/engine/query.py.jinja#L106. When the schema is too large, it's content will be truncated by the OS and the engine will fail to start. The following patch will circumvent the issue by passing the path to the schema instead of it's content: ``` env = os.environ.copy() env.update( # PRISMA_DML=self.dml, RUST_LOG='error', RUST_LOG_FORMAT='json', PRISMA_CLIENT_ENGINE_TYPE='binary', ) if os.environ.get('PRISMA_DML_PATH'): env.update(PRISMA_DML_PATH=os.environ.get('PRISMA_DML_PATH')) else: env.update(PRISMA_DML=self.dml) ``` ## How to reproduce Steps to reproduce the behavior: 1. Generate a prisma.schema file with a large number of tables and fields 2. Generate the prisma client 3. Connect to the database 4. See error: ``` Traceback (most recent call last): File "...", line 239, in <module> app = CashflowLoader(args.process_date) File "...", line 35, in __init__ self.db.connect() File ".../venv/lib/python3.9/site-packages/prisma/client.py", line 3657, in connect self.__engine.connect( File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 110, in connect self.spawn(file, timeout=timeout, datasources=datasources) File ".../venv/lib/python3.9/site-packages/prisma/engine/query.py", line 152, in spawn self.process = subprocess.Popen( File ".../lib/python3.9/subprocess.py", line 951, in __init__ self._execute_child(args, executable, preexec_fn, close_fds, File ".../lib/python3.9/subprocess.py", line 1821, in _execute_child raise child_exception_type(errno_num, err_msg, err_filename) OSError: [Errno 7] Argument list too long: '/tmp/prisma/binaries/engines/efdf9b1183dddfd4258cd181a72125755215ab7b/prisma-query-engine-debian-openssl-1.1.x' ``` ## Expected behavior No error should be reported and the connexion should be established with the database. ## Prisma information Create a large schema file No specific queries required ## Environment & setup - OS: Windows 11, Ubuntu 20.10 - Database: MariaDB - Python version: 3.9.7 - Prisma version: ``` prisma : 3.13.0 prisma client python : 0.6.6 platform : debian-openssl-1.1.x engines : efdf9b1183dddfd4258cd181a72125755215ab7b install path : /mnt/c/Users/andrew/OneDrive/Git_Repos/e22/Benzaiten_repos/benzaiten-processes/venv/lib/python3.9/site-packages/prisma installed extras : [] ```
process
argument list too long when connecting to a database with a large schema bug description the query engine cannot spawn the query engine child process when the schema is too large the error is coming from that line when the schema is too large it s content will be truncated by the os and the engine will fail to start the following patch will circumvent the issue by passing the path to the schema instead of it s content env os environ copy env update prisma dml self dml rust log error rust log format json prisma client engine type binary if os environ get prisma dml path env update prisma dml path os environ get prisma dml path else env update prisma dml self dml how to reproduce steps to reproduce the behavior generate a prisma schema file with a large number of tables and fields generate the prisma client connect to the database see error traceback most recent call last file line in app cashflowloader args process date file line in init self db connect file venv lib site packages prisma client py line in connect self engine connect file venv lib site packages prisma engine query py line in connect self spawn file timeout timeout datasources datasources file venv lib site packages prisma engine query py line in spawn self process subprocess popen file lib subprocess py line in init self execute child args executable preexec fn close fds file lib subprocess py line in execute child raise child exception type errno num err msg err filename oserror argument list too long tmp prisma binaries engines prisma query engine debian openssl x expected behavior no error should be reported and the connexion should be established with the database prisma information create a large schema file no specific queries required environment setup os windows ubuntu database mariadb python version prisma version prisma prisma client python platform debian openssl x engines install path mnt c users andrew onedrive git repos benzaiten repos benzaiten processes venv lib site packages prisma installed extras
1
15,229
19,100,935,004
IssuesEvent
2021-11-29 22:26:47
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Bazel wants full dependency graph for incompatible targets - why aren't they just skipped?
type: support / not a bug (process) team-Configurability untriaged
### Description of the problem / feature request: Hi! I have a target that is only compatible with the `aarch64` platform: ``` cc_binary( name = "foo", target_compatible_with = ["@platforms//cpu:aarch64"], deps = [ "//foo:bar", ] ) ``` Now, the `bar` dependency comes in two flavors, so I have an alias: ``` alias( name = "bar", actual = select({ # Simplified syntax here "aarch64_linux" : ":bar_linux", "aarch64_qnx": ":bar_qnx", }) ) ``` Now, my problem is that when I build for `x86` using wildcard, Bazel complains that it cannot find any `bar` for x86. And that's of course expected - there's no `bar` or x86! However why is Bazel complaining here? I specified that this target is incompatible with `x86`, so it shouldn't need to analyze it. Why is it doing that? If I remove the `bar` dependency, Bazel just skips the target and exits gracefully without errors. I could solve this problem adding a default condition for `bar`, but I don't want that. It will silence a useful warning. If I ever want to support a 3rd `aarch64` version, the default condition will silence the warning without prompting me to add a 3rd `aarch64` version of `bar`. I also don't want to add a `select` in the `deps` field of the `cc_binary` either - it's redundant and adds noise. I already specify that the target is only compatible with`aarch64`. I don't need to add the `deps` only for `aarch64`. Is this expected? Is there any other way I could solve this problem? Thanks! ### Feature requests: what underlying problem are you trying to solve with this feature? I want Bazel to ignore incompatible targets, without it trying to find its dependencies (which may not exist at all for an incompatible platform) ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. See above. ### What operating system are you running Bazel on? Ubuntu 18.04 ### What's the output of `bazel info release`? release 5.0.0-pre.20210907.1 ### Have you found anything relevant by searching the web? Nothing on the web. Asked in [bazel-discuss](https://groups.google.com/g/bazel-discuss/c/NNlN2Jl_bWI) without answer.
1.0
Bazel wants full dependency graph for incompatible targets - why aren't they just skipped? - ### Description of the problem / feature request: Hi! I have a target that is only compatible with the `aarch64` platform: ``` cc_binary( name = "foo", target_compatible_with = ["@platforms//cpu:aarch64"], deps = [ "//foo:bar", ] ) ``` Now, the `bar` dependency comes in two flavors, so I have an alias: ``` alias( name = "bar", actual = select({ # Simplified syntax here "aarch64_linux" : ":bar_linux", "aarch64_qnx": ":bar_qnx", }) ) ``` Now, my problem is that when I build for `x86` using wildcard, Bazel complains that it cannot find any `bar` for x86. And that's of course expected - there's no `bar` or x86! However why is Bazel complaining here? I specified that this target is incompatible with `x86`, so it shouldn't need to analyze it. Why is it doing that? If I remove the `bar` dependency, Bazel just skips the target and exits gracefully without errors. I could solve this problem adding a default condition for `bar`, but I don't want that. It will silence a useful warning. If I ever want to support a 3rd `aarch64` version, the default condition will silence the warning without prompting me to add a 3rd `aarch64` version of `bar`. I also don't want to add a `select` in the `deps` field of the `cc_binary` either - it's redundant and adds noise. I already specify that the target is only compatible with`aarch64`. I don't need to add the `deps` only for `aarch64`. Is this expected? Is there any other way I could solve this problem? Thanks! ### Feature requests: what underlying problem are you trying to solve with this feature? I want Bazel to ignore incompatible targets, without it trying to find its dependencies (which may not exist at all for an incompatible platform) ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. See above. ### What operating system are you running Bazel on? Ubuntu 18.04 ### What's the output of `bazel info release`? release 5.0.0-pre.20210907.1 ### Have you found anything relevant by searching the web? Nothing on the web. Asked in [bazel-discuss](https://groups.google.com/g/bazel-discuss/c/NNlN2Jl_bWI) without answer.
process
bazel wants full dependency graph for incompatible targets why aren t they just skipped description of the problem feature request hi i have a target that is only compatible with the platform cc binary name foo target compatible with deps now the bar dependency comes in two flavors so i have an alias alias name bar actual select simplified syntax here linux bar linux qnx bar qnx now my problem is that when i build for using wildcard bazel complains that it cannot find any bar for and that s of course expected there s no bar or however why is bazel complaining here i specified that this target is incompatible with so it shouldn t need to analyze it why is it doing that if i remove the bar dependency bazel just skips the target and exits gracefully without errors i could solve this problem adding a default condition for bar but i don t want that it will silence a useful warning if i ever want to support a version the default condition will silence the warning without prompting me to add a version of bar i also don t want to add a select in the deps field of the cc binary either it s redundant and adds noise i already specify that the target is only compatible with i don t need to add the deps only for is this expected is there any other way i could solve this problem thanks feature requests what underlying problem are you trying to solve with this feature i want bazel to ignore incompatible targets without it trying to find its dependencies which may not exist at all for an incompatible platform bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible see above what operating system are you running bazel on ubuntu what s the output of bazel info release release pre have you found anything relevant by searching the web nothing on the web asked in without answer
1
415,403
12,128,983,184
IssuesEvent
2020-04-22 21:33:16
microsoftgraph/microsoft-graph-toolkit
https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit
closed
[BUG] In person initials non English letters get ignored
Priority: 0 State: Completed State: Started bug work-in-progress
<!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> **Describe the bug** We filter out non A-Z characters in the initials, should we not use more intelligent regex to filter out bad characters (assumedly numbers, and punctuation etc.). A quick google suggests \p{L} might work. **To Reproduce** Steps to reproduce the behavior: 1. Use the person card with a user whose first or second name starts with initials that aren't a-z. E.g Ægir Davidsson and doesn't have an avatar 2. View initials image 3. See initials as just D 4. JSFiddle: https://jsfiddle.net/robpethi/a08dwyh7/3/ <!-- if possible, please link to a GitHub project that reproduces the issue --> **Expected behavior** Expected initials to be ÆD **Environment (please complete the following information):** - Framework react - Context Teams - Version Latest
1.0
[BUG] In person initials non English letters get ignored - <!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> **Describe the bug** We filter out non A-Z characters in the initials, should we not use more intelligent regex to filter out bad characters (assumedly numbers, and punctuation etc.). A quick google suggests \p{L} might work. **To Reproduce** Steps to reproduce the behavior: 1. Use the person card with a user whose first or second name starts with initials that aren't a-z. E.g Ægir Davidsson and doesn't have an avatar 2. View initials image 3. See initials as just D 4. JSFiddle: https://jsfiddle.net/robpethi/a08dwyh7/3/ <!-- if possible, please link to a GitHub project that reproduces the issue --> **Expected behavior** Expected initials to be ÆD **Environment (please complete the following information):** - Framework react - Context Teams - Version Latest
non_process
in person initials non english letters get ignored describe the bug we filter out non a z characters in the initials should we not use more intelligent regex to filter out bad characters assumedly numbers and punctuation etc a quick google suggests p l might work to reproduce steps to reproduce the behavior use the person card with a user whose first or second name starts with initials that aren t a z e g ægir davidsson and doesn t have an avatar view initials image see initials as just d jsfiddle expected behavior expected initials to be æd environment please complete the following information framework react context teams version latest
0
192,576
22,215,973,194
IssuesEvent
2022-06-08 01:42:43
artsking/linux-4.1.15
https://api.github.com/repos/artsking/linux-4.1.15
reopened
CVE-2016-4470 (Medium) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2016-4470 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command. <p>Publish Date: 2016-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4470>CVE-2016-4470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p> <p>Release Date: 2016-06-27</p> <p>Fix Resolution: v4.7-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-4470 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-4470 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command. <p>Publish Date: 2016-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4470>CVE-2016-4470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p> <p>Release Date: 2016-06-27</p> <p>Fix Resolution: v4.7-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the key reject and link function in security keys key c in the linux kernel through does not ensure that a certain data structure is initialized which allows local users to cause a denial of service system crash via vectors involving a crafted keyctl command publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,691
26,045,591,858
IssuesEvent
2022-12-22 14:08:53
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Allow user to display meaning of exit status in `verdi process list`
topic/verdi type/accepted feature topic/processes
### Is your feature request related to a problem? Please describe `verdi process list` presents information about why calculations fail: ``` $ verdi process list -a -L PwCalculation PK Created Process label Process State Process status ---- --------- --------------- ---------------- --------------------- 194 21h ago PwCalculation ☠ Killed Killed by parent<186> 243 20h ago PwCalculation ⏹ Finished [410] 280 20h ago PwCalculation ☠ Killed Killed by parent<263> 287 20h ago PwCalculation ☠ Killed Killed by parent<282> 733 18h ago PwCalculation ⏹ Finished [0] 741 18h ago PwCalculation ⏹ Finished [400] 785 16h ago PwCalculation ⏹ Finished [0] 790 16h ago PwCalculation ⏹ Finished [312] 834 16h ago PwCalculation ⏹ Finished [340] 837 16h ago PwCalculation ⏹ Finished [0] 856 13h ago PwCalculation ⏹ Finished [305] 873 13h ago PwCalculation ⏹ Finished [0] 887 12h ago PwCalculation ⏹ Finished [305] 903 12h ago PwCalculation ⏹ Finished [305] ``` Some of it is self-explanatory ("Killed"), but the exit codes are not (and they depend on the plugin, so it's not enough to "learn them once"). In order to figure out what they mean, a new user has to go through the reports of the individual calculations. ### Describe the solution you'd like It would be great if a user could request to print the message associated with the exit status as a last column in `verdi process list`. @sphuber What do you think?
1.0
Allow user to display meaning of exit status in `verdi process list` - ### Is your feature request related to a problem? Please describe `verdi process list` presents information about why calculations fail: ``` $ verdi process list -a -L PwCalculation PK Created Process label Process State Process status ---- --------- --------------- ---------------- --------------------- 194 21h ago PwCalculation ☠ Killed Killed by parent<186> 243 20h ago PwCalculation ⏹ Finished [410] 280 20h ago PwCalculation ☠ Killed Killed by parent<263> 287 20h ago PwCalculation ☠ Killed Killed by parent<282> 733 18h ago PwCalculation ⏹ Finished [0] 741 18h ago PwCalculation ⏹ Finished [400] 785 16h ago PwCalculation ⏹ Finished [0] 790 16h ago PwCalculation ⏹ Finished [312] 834 16h ago PwCalculation ⏹ Finished [340] 837 16h ago PwCalculation ⏹ Finished [0] 856 13h ago PwCalculation ⏹ Finished [305] 873 13h ago PwCalculation ⏹ Finished [0] 887 12h ago PwCalculation ⏹ Finished [305] 903 12h ago PwCalculation ⏹ Finished [305] ``` Some of it is self-explanatory ("Killed"), but the exit codes are not (and they depend on the plugin, so it's not enough to "learn them once"). In order to figure out what they mean, a new user has to go through the reports of the individual calculations. ### Describe the solution you'd like It would be great if a user could request to print the message associated with the exit status as a last column in `verdi process list`. @sphuber What do you think?
process
allow user to display meaning of exit status in verdi process list is your feature request related to a problem please describe verdi process list presents information about why calculations fail verdi process list a l pwcalculation pk created process label process state process status ago pwcalculation ☠ killed killed by parent ago pwcalculation ⏹ finished ago pwcalculation ☠ killed killed by parent ago pwcalculation ☠ killed killed by parent ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished ago pwcalculation ⏹ finished some of it is self explanatory killed but the exit codes are not and they depend on the plugin so it s not enough to learn them once in order to figure out what they mean a new user has to go through the reports of the individual calculations describe the solution you d like it would be great if a user could request to print the message associated with the exit status as a last column in verdi process list sphuber what do you think
1
20,357
27,014,590,628
IssuesEvent
2023-02-10 18:08:47
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
Warning: a recent release failed
type: process
The following release PRs may have failed: * #1464 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1414 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1366 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1064 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #998 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #992 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #984 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #1464 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1414 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1366 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #1064 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #998 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #992 - The release job is 'autorelease: tagged', but expected 'autorelease: published'. * #984 - The release job is 'autorelease: tagged', but expected 'autorelease: published'.
process
warning a recent release failed the following release prs may have failed the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published the release job is autorelease tagged but expected autorelease published
1
3,481
6,553,545,824
IssuesEvent
2017-09-05 23:18:14
w3c/w3process
https://api.github.com/repos/w3c/w3process
closed
ToC/document mismatches
Editorial improvements Process2018Candidate
Section in the expanded table of contents, but not in the document: * 6.2.2.1 Substantive Change Sections in the document, but not in the expanded table of contents: * 2.1.2.1 Membership Consortia * 2.1.2.2 Related Members * 2.1.3.1 Advisory Committee Mailing Lists * 2.1.3.2 Advisory Committee Meetings * 5.2.1.1 Member Representative in a Working Group * 5.2.1.2 Member Representative in an Interest Group * 5.2.1.3 Invited Expert in a Working Group * 5.2.1.4 Invited Expert in an Interest Group * 5.2.1.5 Team Representative in a Working Group * 5.2.1.6 Team Representative in an Interest Group Title mismatches between ToC and document: * 6.2.5 "Classes of Changes to a Recommendation" vs. "Classes of Changes" * 10.4 "Rejection of a Submission Request" vs. "Rejection of a Submission Request, and Submission Appeals" Reference mismatch: In [6.7.1 Errata Management](https://www.w3.org/2017/Process-20170301/#errata), there's a reference to "7.2.5 Classes of Changes", but it should be "6.2.5 Classes of Changes". Version checked: Process Document 1 March 2017 (I haven't checked the latest Editor's version, but I think it would suffer from at least some of the problems above.) Automatically numbered and generated ToC/references would be great.
1.0
ToC/document mismatches - Section in the expanded table of contents, but not in the document: * 6.2.2.1 Substantive Change Sections in the document, but not in the expanded table of contents: * 2.1.2.1 Membership Consortia * 2.1.2.2 Related Members * 2.1.3.1 Advisory Committee Mailing Lists * 2.1.3.2 Advisory Committee Meetings * 5.2.1.1 Member Representative in a Working Group * 5.2.1.2 Member Representative in an Interest Group * 5.2.1.3 Invited Expert in a Working Group * 5.2.1.4 Invited Expert in an Interest Group * 5.2.1.5 Team Representative in a Working Group * 5.2.1.6 Team Representative in an Interest Group Title mismatches between ToC and document: * 6.2.5 "Classes of Changes to a Recommendation" vs. "Classes of Changes" * 10.4 "Rejection of a Submission Request" vs. "Rejection of a Submission Request, and Submission Appeals" Reference mismatch: In [6.7.1 Errata Management](https://www.w3.org/2017/Process-20170301/#errata), there's a reference to "7.2.5 Classes of Changes", but it should be "6.2.5 Classes of Changes". Version checked: Process Document 1 March 2017 (I haven't checked the latest Editor's version, but I think it would suffer from at least some of the problems above.) Automatically numbered and generated ToC/references would be great.
process
toc document mismatches section in the expanded table of contents but not in the document substantive change sections in the document but not in the expanded table of contents membership consortia related members advisory committee mailing lists advisory committee meetings member representative in a working group member representative in an interest group invited expert in a working group invited expert in an interest group team representative in a working group team representative in an interest group title mismatches between toc and document classes of changes to a recommendation vs classes of changes rejection of a submission request vs rejection of a submission request and submission appeals reference mismatch in there s a reference to classes of changes but it should be classes of changes version checked process document march i haven t checked the latest editor s version but i think it would suffer from at least some of the problems above automatically numbered and generated toc references would be great
1
21,532
3,517,357,176
IssuesEvent
2016-01-12 06:59:43
sanderv32/virtualboxserverservice
https://api.github.com/repos/sanderv32/virtualboxserverservice
closed
Installer hangs when trying to re-install
Component-Scripts Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Run the installer once (successfully) 2. Run it again (try to reinstall) => Installer hangs on "Creating scheduled task". Can't press cancel or even the window. I tried this on 64bit Windows 7 Ultimate with VBoxService-setup_2.2.0.39135-BETA.exe. ``` Original issue reported on code.google.com by `jarno.el...@gmail.com` on 4 Dec 2010 at 4:34
1.0
Installer hangs when trying to re-install - ``` What steps will reproduce the problem? 1. Run the installer once (successfully) 2. Run it again (try to reinstall) => Installer hangs on "Creating scheduled task". Can't press cancel or even the window. I tried this on 64bit Windows 7 Ultimate with VBoxService-setup_2.2.0.39135-BETA.exe. ``` Original issue reported on code.google.com by `jarno.el...@gmail.com` on 4 Dec 2010 at 4:34
non_process
installer hangs when trying to re install what steps will reproduce the problem run the installer once successfully run it again try to reinstall installer hangs on creating scheduled task can t press cancel or even the window i tried this on windows ultimate with vboxservice setup beta exe original issue reported on code google com by jarno el gmail com on dec at
0
102,527
22,032,848,378
IssuesEvent
2022-05-28 05:21:45
NicoVIII/Andromeda-for-GOG
https://api.github.com/repos/NicoVIII/Andromeda-for-GOG
closed
Add CodeTour
code
With the CodeTour extension for VScode it is possible to record Tours with comments and explain the structure of the project. https://marketplace.visualstudio.com/items?itemName=vsls-contrib.codetour This is a nice approach to make code easier to understand for new developers, so I want to try that.
1.0
Add CodeTour - With the CodeTour extension for VScode it is possible to record Tours with comments and explain the structure of the project. https://marketplace.visualstudio.com/items?itemName=vsls-contrib.codetour This is a nice approach to make code easier to understand for new developers, so I want to try that.
non_process
add codetour with the codetour extension for vscode it is possible to record tours with comments and explain the structure of the project this is a nice approach to make code easier to understand for new developers so i want to try that
0
34,449
12,288,116,359
IssuesEvent
2020-05-09 15:20:00
Zymergen/hubot-docker
https://api.github.com/repos/Zymergen/hubot-docker
opened
WS-2018-0209 (Medium) detected in morgan-1.6.1.tgz
security vulnerability
## WS-2018-0209 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary> <p>HTTP request logger middleware for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/hubot-docker/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/hubot-docker/node_modules/morgan/package.json</p> <p> Dependency Hierarchy: - hubot-2.19.0.tgz (Root Library) - express-3.21.2.tgz - connect-2.30.2.tgz - :x: **morgan-1.6.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Zymergen/hubot-docker/commit/07953cb6bb385a84410fb77bc2c3d2ff16dee495">07953cb6bb385a84410fb77bc2c3d2ff16dee495</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack. <p>Publish Date: 2018-11-25 <p>URL: <a href=>WS-2018-0209</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p> <p>Release Date: 2019-04-08</p> <p>Fix Resolution: 1.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","isTransitiveDependency":true,"dependencyTree":"hubot:2.19.0;express:3.21.2;connect:2.30.2;morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
True
WS-2018-0209 (Medium) detected in morgan-1.6.1.tgz - ## WS-2018-0209 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary> <p>HTTP request logger middleware for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/hubot-docker/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/hubot-docker/node_modules/morgan/package.json</p> <p> Dependency Hierarchy: - hubot-2.19.0.tgz (Root Library) - express-3.21.2.tgz - connect-2.30.2.tgz - :x: **morgan-1.6.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Zymergen/hubot-docker/commit/07953cb6bb385a84410fb77bc2c3d2ff16dee495">07953cb6bb385a84410fb77bc2c3d2ff16dee495</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack. <p>Publish Date: 2018-11-25 <p>URL: <a href=>WS-2018-0209</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p> <p>Release Date: 2019-04-08</p> <p>Fix Resolution: 1.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","isTransitiveDependency":true,"dependencyTree":"hubot:2.19.0;express:3.21.2;connect:2.30.2;morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in morgan tgz ws medium severity vulnerability vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file tmp ws scm hubot docker package json path to vulnerable library tmp ws scm hubot docker node modules morgan package json dependency hierarchy hubot tgz root library express tgz connect tgz x morgan tgz vulnerable library found in head commit a href vulnerability details morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack publish date url ws cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack vulnerabilityurl medium extradata
0
15,682
19,847,795,320
IssuesEvent
2022-01-21 08:53:12
ooi-data/RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample
https://api.github.com/repos/ooi-data/RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:53:11.809404. ## Details Flow name: `RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:53:11.809404. ## Details Flow name: `RS01SBPS-PC01A-4A-CTDPFA103-streamed-ctdpf_optode_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name streamed ctdpf optode sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
2,217
2,599,855,153
IssuesEvent
2015-02-23 12:25:05
lsecities/lsecities-wp-theme
https://api.github.com/repos/lsecities/lsecities-wp-theme
opened
DBUA Award lists - update template
enhancement needs:design
we should display some brief blurb for each award. photos also need to be larger than the current miserly thumbnails...
1.0
DBUA Award lists - update template - we should display some brief blurb for each award. photos also need to be larger than the current miserly thumbnails...
non_process
dbua award lists update template we should display some brief blurb for each award photos also need to be larger than the current miserly thumbnails
0
19,796
26,178,434,019
IssuesEvent
2023-01-02 12:44:07
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Wrong db pull warning with multiSchema preview feature, it says that "These enums were enriched with `@@map` information taken from the previous Prisma schema." but the schema was actually empty.
bug/2-confirmed kind/bug process/candidate topic: introspection-warning topic: re-introspection team/schema topic: multiSchema
### Bug description The following warnings are confusing ``` // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" ``` Here we are in a re-introspection case but there is no datamodel, only a generator and a datasource. The warning suggests that some information was taken from this empty datamodel which is impossible. It's probably the logic in re-introspection in the engine that needs to be changed. ### How to reproduce - create a database with `CREATE DATABASE "reintroMultiSchema";` - import / execute the following SQL ```sql -- CreateSchema CREATE SCHEMA IF NOT EXISTS "base"; -- CreateSchema CREATE SCHEMA IF NOT EXISTS "transactional"; -- CreateTable CREATE TABLE "base"."User" ( "id" TEXT NOT NULL, "email" TEXT NOT NULL, CONSTRAINT "User_pkey" PRIMARY KEY ("id") ); -- CreateTable CREATE TABLE "transactional"."Post" ( "id" TEXT NOT NULL, "title" TEXT NOT NULL, "authorId" TEXT NOT NULL, CONSTRAINT "Post_pkey" PRIMARY KEY ("id") ); -- AddForeignKey ALTER TABLE "transactional"."Post" ADD CONSTRAINT "Post_authorId_fkey" FOREIGN KEY ("authorId") REFERENCES "base"."User"("id") ON DELETE RESTRICT ON UPDATE CASCADE; CREATE TYPE "base"."status" as enum ('ON','OFF'); -- -- tables names are renamed so they are identical -- Original names are User and Post -- Errors with Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field. -- https://github.com/prisma/prisma/issues/15800 -- -- CreateTable CREATE TABLE "base"."some_table" ( "id" TEXT NOT NULL, "email" TEXT NOT NULL, CONSTRAINT "User_pkey2" PRIMARY KEY ("id") ); -- CreateTable CREATE TABLE "transactional"."some_table" ( "id" TEXT NOT NULL, "title" TEXT NOT NULL, "authorId" TEXT NOT NULL, CONSTRAINT "Post_pkey2" PRIMARY KEY ("id") ); -- AddForeignKey ALTER TABLE "transactional"."some_table" ADD CONSTRAINT "Post_authorId_fkey2" FOREIGN KEY ("authorId") REFERENCES "base"."some_table"("id") ON DELETE RESTRICT ON UPDATE CASCADE; CREATE TYPE "transactional"."status" as enum ('ON','OFF'); ``` - run `DATABASE_URL="postgres://prisma:prisma@localhost:5432/reintroMultiSchema" npx prisma@dev db pull --print --schema schema.prisma` Result ``` generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["base", "transactional"] } model User { id String @id email String Post Post[] @@schema("base") } model base_some_table { id String @id(map: "User_pkey2") email String some_table transactional_some_table[] @@map("some_table") @@schema("base") } model Post { id String @id title String authorId String User User @relation(fields: [authorId], references: [id]) @@schema("transactional") } model transactional_some_table { id String @id(map: "Post_pkey2") title String authorId String some_table base_some_table @relation(fields: [authorId], references: [id], map: "Post_authorId_fkey2") @@map("some_table") @@schema("transactional") } enum base_status { ON OFF @@map("status") @@schema("base") } enum transactional_status { ON OFF @@map("status") @@schema("transactional") } // introspectionSchemaVersion: NonPrisma // *** WARNING *** // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" // // These models and enums were renamed due to their names being duplicates in the Prisma Schema Language. // Code 20 // [ // { // "type": "Enum", // "name": "base_status" // }, // { // "type": "Enum", // "name": "transactional_status" // }, // { // "type": "Model", // "name": "base_some_table" // }, // { // "type": "Model", // "name": "transactional_some_table" // } // ] // ``` ### Expected behavior These warnings should not be outputted ``` // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" ``` ### Prisma information ```prisma datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["base", "transactional"] } generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] } ``` ### Environment & setup - OS: macOS - Database: PostgreSQL - Node.js version: NA ### Prisma Version ``` prisma : 4.9.0-dev.13 @prisma/client : Not found Current platform : darwin-arm64 Query Engine (Node-API) : libquery-engine c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node) Migration Engine : migration-engine-cli c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/migration-engine-darwin-arm64) Introspection Engine : introspection-core c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/introspection-engine-darwin-arm64) Format Binary : prisma-fmt c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/prisma-fmt-darwin-arm64) Format Wasm : @prisma/prisma-fmt-wasm 4.9.0-6.c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 Default Engines Hash : c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 Studio : 0.479.0 Preview Features : multiSchema ```
1.0
Wrong db pull warning with multiSchema preview feature, it says that "These enums were enriched with `@@map` information taken from the previous Prisma schema." but the schema was actually empty. - ### Bug description The following warnings are confusing ``` // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" ``` Here we are in a re-introspection case but there is no datamodel, only a generator and a datasource. The warning suggests that some information was taken from this empty datamodel which is impossible. It's probably the logic in re-introspection in the engine that needs to be changed. ### How to reproduce - create a database with `CREATE DATABASE "reintroMultiSchema";` - import / execute the following SQL ```sql -- CreateSchema CREATE SCHEMA IF NOT EXISTS "base"; -- CreateSchema CREATE SCHEMA IF NOT EXISTS "transactional"; -- CreateTable CREATE TABLE "base"."User" ( "id" TEXT NOT NULL, "email" TEXT NOT NULL, CONSTRAINT "User_pkey" PRIMARY KEY ("id") ); -- CreateTable CREATE TABLE "transactional"."Post" ( "id" TEXT NOT NULL, "title" TEXT NOT NULL, "authorId" TEXT NOT NULL, CONSTRAINT "Post_pkey" PRIMARY KEY ("id") ); -- AddForeignKey ALTER TABLE "transactional"."Post" ADD CONSTRAINT "Post_authorId_fkey" FOREIGN KEY ("authorId") REFERENCES "base"."User"("id") ON DELETE RESTRICT ON UPDATE CASCADE; CREATE TYPE "base"."status" as enum ('ON','OFF'); -- -- tables names are renamed so they are identical -- Original names are User and Post -- Errors with Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field. -- https://github.com/prisma/prisma/issues/15800 -- -- CreateTable CREATE TABLE "base"."some_table" ( "id" TEXT NOT NULL, "email" TEXT NOT NULL, CONSTRAINT "User_pkey2" PRIMARY KEY ("id") ); -- CreateTable CREATE TABLE "transactional"."some_table" ( "id" TEXT NOT NULL, "title" TEXT NOT NULL, "authorId" TEXT NOT NULL, CONSTRAINT "Post_pkey2" PRIMARY KEY ("id") ); -- AddForeignKey ALTER TABLE "transactional"."some_table" ADD CONSTRAINT "Post_authorId_fkey2" FOREIGN KEY ("authorId") REFERENCES "base"."some_table"("id") ON DELETE RESTRICT ON UPDATE CASCADE; CREATE TYPE "transactional"."status" as enum ('ON','OFF'); ``` - run `DATABASE_URL="postgres://prisma:prisma@localhost:5432/reintroMultiSchema" npx prisma@dev db pull --print --schema schema.prisma` Result ``` generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["base", "transactional"] } model User { id String @id email String Post Post[] @@schema("base") } model base_some_table { id String @id(map: "User_pkey2") email String some_table transactional_some_table[] @@map("some_table") @@schema("base") } model Post { id String @id title String authorId String User User @relation(fields: [authorId], references: [id]) @@schema("transactional") } model transactional_some_table { id String @id(map: "Post_pkey2") title String authorId String some_table base_some_table @relation(fields: [authorId], references: [id], map: "Post_authorId_fkey2") @@map("some_table") @@schema("transactional") } enum base_status { ON OFF @@map("status") @@schema("base") } enum transactional_status { ON OFF @@map("status") @@schema("transactional") } // introspectionSchemaVersion: NonPrisma // *** WARNING *** // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" // // These models and enums were renamed due to their names being duplicates in the Prisma Schema Language. // Code 20 // [ // { // "type": "Enum", // "name": "base_status" // }, // { // "type": "Enum", // "name": "transactional_status" // }, // { // "type": "Model", // "name": "base_some_table" // }, // { // "type": "Model", // "name": "transactional_some_table" // } // ] // ``` ### Expected behavior These warnings should not be outputted ``` // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "base_status" // // These enums were enriched with `@@map` information taken from the previous Prisma schema. // - Enum "transactional_status" ``` ### Prisma information ```prisma datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["base", "transactional"] } generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] } ``` ### Environment & setup - OS: macOS - Database: PostgreSQL - Node.js version: NA ### Prisma Version ``` prisma : 4.9.0-dev.13 @prisma/client : Not found Current platform : darwin-arm64 Query Engine (Node-API) : libquery-engine c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node) Migration Engine : migration-engine-cli c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/migration-engine-darwin-arm64) Introspection Engine : introspection-core c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/introspection-engine-darwin-arm64) Format Binary : prisma-fmt c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 (at ../../.npm/_npx/1eb32a0be9e7d512/node_modules/@prisma/engines/prisma-fmt-darwin-arm64) Format Wasm : @prisma/prisma-fmt-wasm 4.9.0-6.c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 Default Engines Hash : c417b3d7f4d3ba648bde7483f42ddf7a1f0cbd35 Studio : 0.479.0 Preview Features : multiSchema ```
process
wrong db pull warning with multischema preview feature it says that these enums were enriched with map information taken from the previous prisma schema but the schema was actually empty bug description the following warnings are confusing these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status here we are in a re introspection case but there is no datamodel only a generator and a datasource the warning suggests that some information was taken from this empty datamodel which is impossible it s probably the logic in re introspection in the engine that needs to be changed how to reproduce create a database with create database reintromultischema import execute the following sql sql createschema create schema if not exists base createschema create schema if not exists transactional createtable create table base user id text not null email text not null constraint user pkey primary key id createtable create table transactional post id text not null title text not null authorid text not null constraint post pkey primary key id addforeignkey alter table transactional post add constraint post authorid fkey foreign key authorid references base user id on delete restrict on update cascade create type base status as enum on off tables names are renamed so they are identical original names are user and post errors with error every relationinfo should have a complementary relationinfo on the opposite relation field createtable create table base some table id text not null email text not null constraint user primary key id createtable create table transactional some table id text not null title text not null authorid text not null constraint post primary key id addforeignkey alter table transactional some table add constraint post authorid foreign key authorid references base some table id on delete restrict on update cascade create type transactional status as enum on off run database url postgres prisma prisma localhost reintromultischema npx prisma dev db pull print schema schema prisma result generator client provider prisma client js previewfeatures datasource db provider postgresql url env database url schemas model user id string id email string post post schema base model base some table id string id map user email string some table transactional some table map some table schema base model post id string id title string authorid string user user relation fields references schema transactional model transactional some table id string id map post title string authorid string some table base some table relation fields references map post authorid map some table schema transactional enum base status on off map status schema base enum transactional status on off map status schema transactional introspectionschemaversion nonprisma warning these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status these models and enums were renamed due to their names being duplicates in the prisma schema language code type enum name base status type enum name transactional status type model name base some table type model name transactional some table expected behavior these warnings should not be outputted these enums were enriched with map information taken from the previous prisma schema enum base status these enums were enriched with map information taken from the previous prisma schema enum transactional status prisma information prisma datasource db provider postgresql url env database url schemas generator client provider prisma client js previewfeatures environment setup os macos database postgresql node js version na prisma version prisma dev prisma client not found current platform darwin query engine node api libquery engine at npm npx node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at npm npx node modules prisma engines migration engine darwin introspection engine introspection core at npm npx node modules prisma engines introspection engine darwin format binary prisma fmt at npm npx node modules prisma engines prisma fmt darwin format wasm prisma prisma fmt wasm default engines hash studio preview features multischema
1
386,956
11,453,716,869
IssuesEvent
2020-02-06 15:51:36
easystats/parameters
https://api.github.com/repos/easystats/parameters
closed
Next CRAN release
high priority :running_man:
Due to changes in the latest R devel I have to update the performance and sjstats packages. however since parameters has new functions that are required for the latest performance function I think of submitting parameters first. so are there any urgent features or bugs that needs to be fixed before submission? I'm planning to submit next week.
1.0
Next CRAN release - Due to changes in the latest R devel I have to update the performance and sjstats packages. however since parameters has new functions that are required for the latest performance function I think of submitting parameters first. so are there any urgent features or bugs that needs to be fixed before submission? I'm planning to submit next week.
non_process
next cran release due to changes in the latest r devel i have to update the performance and sjstats packages however since parameters has new functions that are required for the latest performance function i think of submitting parameters first so are there any urgent features or bugs that needs to be fixed before submission i m planning to submit next week
0
134,395
10,906,262,527
IssuesEvent
2019-11-20 12:34:22
Laravel-Backpack/CRUD
https://api.github.com/repos/Laravel-Backpack/CRUD
closed
[4.0.7] select2_ajax filter breaks "Remove filters" button
testing or needs confirmation working on it
# Bug report ### What I did 1 - Add a select2_ajax filter 2 - Filter the results 3 - Click in the "Remove filters" button ### What I expected to happen The filters should be removed ### What happened The filter is removed from the interface, but the list return zero results: ![image](https://i.imgur.com/UOqAQNw.gif) ### What I've already tried to fix it What I can tell is that the filter key is still present when we remove the filter, but it has no value. ![image](https://user-images.githubusercontent.com/4185886/66430076-c1d61880-e9ef-11e9-8f6e-00167ee05790.png) To avoid the empty results I added a check in the filter function: ![image](https://user-images.githubusercontent.com/4185886/66430400-6ce6d200-e9f0-11e9-8f8c-f681cfc151c4.png) ### Backpack, Laravel, PHP, DB version ### PHP VERSION: PHP 7.2.22 (cli) (built: Sep 8 2019 15:19:13) ( NTS ) ### LARAVEL VERSION: laravel/framework v6.1.0 ### BACKPACK VERSION: backpack/crud 4.0.7 backpack/generators 2.0.4 backpack/permissionmanager 5.0.1
1.0
[4.0.7] select2_ajax filter breaks "Remove filters" button - # Bug report ### What I did 1 - Add a select2_ajax filter 2 - Filter the results 3 - Click in the "Remove filters" button ### What I expected to happen The filters should be removed ### What happened The filter is removed from the interface, but the list return zero results: ![image](https://i.imgur.com/UOqAQNw.gif) ### What I've already tried to fix it What I can tell is that the filter key is still present when we remove the filter, but it has no value. ![image](https://user-images.githubusercontent.com/4185886/66430076-c1d61880-e9ef-11e9-8f6e-00167ee05790.png) To avoid the empty results I added a check in the filter function: ![image](https://user-images.githubusercontent.com/4185886/66430400-6ce6d200-e9f0-11e9-8f8c-f681cfc151c4.png) ### Backpack, Laravel, PHP, DB version ### PHP VERSION: PHP 7.2.22 (cli) (built: Sep 8 2019 15:19:13) ( NTS ) ### LARAVEL VERSION: laravel/framework v6.1.0 ### BACKPACK VERSION: backpack/crud 4.0.7 backpack/generators 2.0.4 backpack/permissionmanager 5.0.1
non_process
ajax filter breaks remove filters button bug report what i did add a ajax filter filter the results click in the remove filters button what i expected to happen the filters should be removed what happened the filter is removed from the interface but the list return zero results what i ve already tried to fix it what i can tell is that the filter key is still present when we remove the filter but it has no value to avoid the empty results i added a check in the filter function backpack laravel php db version php version php cli built sep nts laravel version laravel framework backpack version backpack crud backpack generators backpack permissionmanager
0
83,067
23,959,015,857
IssuesEvent
2022-09-12 17:19:11
xamarin/xamarin-android
https://api.github.com/repos/xamarin/xamarin-android
closed
Generate OpenTK JCW just once and then merge them to all the API levels
Area: xamarin-android Build
Instead of generating them per level. ``` GenerateJavaCallableWrappers: mono --debug=casts "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild/Xamarin/Android/jcw-gen.exe" -v10 -o "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/jcw/src" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0/Facades" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/OpenTK-1.0.dll" "javac" -source 1.6 -target 1.6 -d "obj/Debug/jcw/bin" -bootclasspath "/Users/builder/android-toolchain/sdk/platforms/android-21/android.jar":"/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" @obj/Debug/jcw/classes.txt Note: Some input files use unchecked or unsafe operations. Note: Recompile with -Xlint:unchecked for details. "jar" uf "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" -C "obj/Debug/jcw/bin" . _GenerateMonoAndroidDex16: "/Users/builder/android-toolchain/sdk/build-tools/28.0.0/dx" --dex --no-strict --output="/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.dex" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" ``` That would mean to run `jcw-gen.exe` once, which will hopefully result in slightly shorter build times.
1.0
Generate OpenTK JCW just once and then merge them to all the API levels - Instead of generating them per level. ``` GenerateJavaCallableWrappers: mono --debug=casts "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild/Xamarin/Android/jcw-gen.exe" -v10 -o "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/jcw/src" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0" -L "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/../v1.0/Facades" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/src/OpenTK-1.0/obj/Debug/OpenTK-1.0.dll" "javac" -source 1.6 -target 1.6 -d "obj/Debug/jcw/bin" -bootclasspath "/Users/builder/android-toolchain/sdk/platforms/android-21/android.jar":"/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" @obj/Debug/jcw/classes.txt Note: Some input files use unchecked or unsafe operations. Note: Recompile with -Xlint:unchecked for details. "jar" uf "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" -C "obj/Debug/jcw/bin" . _GenerateMonoAndroidDex16: "/Users/builder/android-toolchain/sdk/build-tools/28.0.0/dx" --dex --no-strict --output="/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.dex" "/Users/builder/jenkins/workspace/xamarin-android-pr-builder/xamarin-android/bin/Debug/lib/xamarin.android/xbuild-frameworks/MonoAndroid/v5.0/mono.android.jar" ``` That would mean to run `jcw-gen.exe` once, which will hopefully result in slightly shorter build times.
non_process
generate opentk jcw just once and then merge them to all the api levels instead of generating them per level generatejavacallablewrappers mono debug casts users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild xamarin android jcw gen exe o users builder jenkins workspace xamarin android pr builder xamarin android src opentk obj debug jcw src l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid l users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid facades users builder jenkins workspace xamarin android pr builder xamarin android src opentk obj debug opentk dll javac source target d obj debug jcw bin bootclasspath users builder android toolchain sdk platforms android android jar users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar obj debug jcw classes txt note some input files use unchecked or unsafe operations note recompile with xlint unchecked for details jar uf users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar c obj debug jcw bin users builder android toolchain sdk build tools dx dex no strict output users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android dex users builder jenkins workspace xamarin android pr builder xamarin android bin debug lib xamarin android xbuild frameworks monoandroid mono android jar that would mean to run jcw gen exe once which will hopefully result in slightly shorter build times
0
13,361
15,825,459,797
IssuesEvent
2021-04-06 05:49:39
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add the ability to wire-up listeners before starting a child process
child_process feature request
**Is your feature request related to a problem? Please describe.** When registering multiple listeners (callbacks) to the `data` event of a child process's `stdout`, there's no way to get the child process to wait for all the callbacks to be registered before starting. This means there's a window between registering the first listener and the second one in which the first listener might "pull" the first available chunk and when the second listener is registered, it won't receive the first chunk. **A thinned-down example:** ```javascript // Expected behaviour scenario const { exec } = require("child_process"); p = exec("seq 1000"); // This command prints the numbers between 1 and 1000, each in a different line a1 = ''; a2 = ''; p.stdout.on("data", (d) => a1 += d); p.stdout.on("data", (d) => a2 += d); // When the child process completes, a1 and a2 will both contain all the numbers from 1 to 1000 ``` ```javascript // Edge-case scenario const { exec } = require("child_process"); p = exec("seq 1000"); a1 = ''; a2 = ''; p.stdout.on("data", (d) => a1 += d); setTimeout(() => { p.stdout.on("data", (d) => a2 += d); }, 500); // When the child process completes, a1 will contain all the numbers from 1 to 1000 while a2 will remain an empty string ``` From what I understand from reading the documentation of child_process, when a child process is started nodejs saves it's output in a buffer until a listener is registered (either by directly binding to the 'data' event or by `pipe()`ing stdout to a writable stream). This behaviour creates two potential problems: 1. A second listener might not get the same data as the first one. 2. the child process might output more data than the buffer can contain before any data can be processed. **Describe the solution you'd like** The solution I propose is to allow wiring up all the listeners and pipes before starting the child process. Considering backwards compatibility, I imagine the best way to achieve this is by passing a new option (something like `autostart` that will default to `true`) to the options parameter of `spawn`, `exec` etc., that will make those functions return a `ChildProcess` instance that was not yet started, together with a new `start()` method added to the `ChildProcess` class. **Describe alternatives you've considered** The alternatives as I see them are: 1. Only register a single handler and pass the data around to your multiple destinations. 2. Try to proxy the readable stream to a second one that is already wired up. 3. Try your best to minimize that window and hope for the best.
1.0
Add the ability to wire-up listeners before starting a child process - **Is your feature request related to a problem? Please describe.** When registering multiple listeners (callbacks) to the `data` event of a child process's `stdout`, there's no way to get the child process to wait for all the callbacks to be registered before starting. This means there's a window between registering the first listener and the second one in which the first listener might "pull" the first available chunk and when the second listener is registered, it won't receive the first chunk. **A thinned-down example:** ```javascript // Expected behaviour scenario const { exec } = require("child_process"); p = exec("seq 1000"); // This command prints the numbers between 1 and 1000, each in a different line a1 = ''; a2 = ''; p.stdout.on("data", (d) => a1 += d); p.stdout.on("data", (d) => a2 += d); // When the child process completes, a1 and a2 will both contain all the numbers from 1 to 1000 ``` ```javascript // Edge-case scenario const { exec } = require("child_process"); p = exec("seq 1000"); a1 = ''; a2 = ''; p.stdout.on("data", (d) => a1 += d); setTimeout(() => { p.stdout.on("data", (d) => a2 += d); }, 500); // When the child process completes, a1 will contain all the numbers from 1 to 1000 while a2 will remain an empty string ``` From what I understand from reading the documentation of child_process, when a child process is started nodejs saves it's output in a buffer until a listener is registered (either by directly binding to the 'data' event or by `pipe()`ing stdout to a writable stream). This behaviour creates two potential problems: 1. A second listener might not get the same data as the first one. 2. the child process might output more data than the buffer can contain before any data can be processed. **Describe the solution you'd like** The solution I propose is to allow wiring up all the listeners and pipes before starting the child process. Considering backwards compatibility, I imagine the best way to achieve this is by passing a new option (something like `autostart` that will default to `true`) to the options parameter of `spawn`, `exec` etc., that will make those functions return a `ChildProcess` instance that was not yet started, together with a new `start()` method added to the `ChildProcess` class. **Describe alternatives you've considered** The alternatives as I see them are: 1. Only register a single handler and pass the data around to your multiple destinations. 2. Try to proxy the readable stream to a second one that is already wired up. 3. Try your best to minimize that window and hope for the best.
process
add the ability to wire up listeners before starting a child process is your feature request related to a problem please describe when registering multiple listeners callbacks to the data event of a child process s stdout there s no way to get the child process to wait for all the callbacks to be registered before starting this means there s a window between registering the first listener and the second one in which the first listener might pull the first available chunk and when the second listener is registered it won t receive the first chunk a thinned down example javascript expected behaviour scenario const exec require child process p exec seq this command prints the numbers between and each in a different line p stdout on data d d p stdout on data d d when the child process completes and will both contain all the numbers from to javascript edge case scenario const exec require child process p exec seq p stdout on data d d settimeout p stdout on data d d when the child process completes will contain all the numbers from to while will remain an empty string from what i understand from reading the documentation of child process when a child process is started nodejs saves it s output in a buffer until a listener is registered either by directly binding to the data event or by pipe ing stdout to a writable stream this behaviour creates two potential problems a second listener might not get the same data as the first one the child process might output more data than the buffer can contain before any data can be processed describe the solution you d like the solution i propose is to allow wiring up all the listeners and pipes before starting the child process considering backwards compatibility i imagine the best way to achieve this is by passing a new option something like autostart that will default to true to the options parameter of spawn exec etc that will make those functions return a childprocess instance that was not yet started together with a new start method added to the childprocess class describe alternatives you ve considered the alternatives as i see them are only register a single handler and pass the data around to your multiple destinations try to proxy the readable stream to a second one that is already wired up try your best to minimize that window and hope for the best
1
13,392
15,866,357,527
IssuesEvent
2021-04-08 15:39:53
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Study activities > Value picker response type >App is allowing the user to submit the response without selecting any value
Android Bug P1 Process: Fixed Process: Tested dev
Steps:- 1. Configure an activity with Value Picker based response type in Study builder 2. Open the Android application and login into App 3. Enroll into the study 4. Try to submit a response for Value Picker response type activities and verify AR:- App is allowing the user to submit the response without selecting any value ER:- App should not allow submitting the response until the user selects the value
2.0
[Android] Study activities > Value picker response type >App is allowing the user to submit the response without selecting any value - Steps:- 1. Configure an activity with Value Picker based response type in Study builder 2. Open the Android application and login into App 3. Enroll into the study 4. Try to submit a response for Value Picker response type activities and verify AR:- App is allowing the user to submit the response without selecting any value ER:- App should not allow submitting the response until the user selects the value
process
study activities value picker response type app is allowing the user to submit the response without selecting any value steps configure an activity with value picker based response type in study builder open the android application and login into app enroll into the study try to submit a response for value picker response type activities and verify ar app is allowing the user to submit the response without selecting any value er app should not allow submitting the response until the user selects the value
1
6,708
9,815,605,176
IssuesEvent
2019-06-13 13:01:42
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
closed
Update project README to link to opencue.io
process
**Describe the bug** When we deprecate the Wiki, we'll need to update the project README to link to opencue.io. **To Reproduce** N/A **Expected behavior** There should no longer be links from the main repository to the project Wiki. **Screenshots** N/A **Version Number** N/A **Additional context** N/A
1.0
Update project README to link to opencue.io - **Describe the bug** When we deprecate the Wiki, we'll need to update the project README to link to opencue.io. **To Reproduce** N/A **Expected behavior** There should no longer be links from the main repository to the project Wiki. **Screenshots** N/A **Version Number** N/A **Additional context** N/A
process
update project readme to link to opencue io describe the bug when we deprecate the wiki we ll need to update the project readme to link to opencue io to reproduce n a expected behavior there should no longer be links from the main repository to the project wiki screenshots n a version number n a additional context n a
1
25,761
12,734,975,061
IssuesEvent
2020-06-25 14:40:34
kframework/kore
https://api.github.com/repos/kframework/kore
closed
Comparatively bad performance on some concrete tests
appetite: S investigate performance
When comparing the Haskell backend to the Java backend on KEVM concrete tests (574 tests total), we have: - 392 tests which the Haskell backend finishes faster than the Java backend, - 108 tests which finish within 2x the time of the Java backend (< 2x slowdown), - 36 tests which timeout unexpectedly (see #1348 ), and - 38 tests which take more than 2x the time of the Java backend. This issue concerns the last category of tests, because it indicates that there are more efficient ways to handle those tests which the Java backend is taking advantage of but the Haskell backend is not. This is that list (note that we jump from 1.77x max slowdown in other tests to 2.82x slowdown in this list): ``` 2.82 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_29.json 2.87 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_11.json 2.89 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_8.json 2.90 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_1.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_28.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_9.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/fibbonacci_unrolled.json 2.95 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_23.json 2.96 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_0.json 2.97 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_16.json 2.98 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_25.json 2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_31.json 2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_4.json 3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_21.json 3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_32.json 3.02 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_2.json 3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_10.json 3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_6.json 3.05 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_30.json 3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_12.json 3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_18.json 3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_14.json 3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_20.json 3.08 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_27.json 3.09 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_19.json 3.10 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_24.json 3.11 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_3.json 3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_17.json 3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_22.json 3.13 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_33.json 3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_26.json 3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_5.json 3.16 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_7.json 3.17 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_15.json 3.18 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_13.json 3.69 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpdestBigList.json 4.86 tests/ethereum-tests/VMTests/vmIOandFlowOperations/byte1.json 5.37 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpi_at_the_end.json ```
True
Comparatively bad performance on some concrete tests - When comparing the Haskell backend to the Java backend on KEVM concrete tests (574 tests total), we have: - 392 tests which the Haskell backend finishes faster than the Java backend, - 108 tests which finish within 2x the time of the Java backend (< 2x slowdown), - 36 tests which timeout unexpectedly (see #1348 ), and - 38 tests which take more than 2x the time of the Java backend. This issue concerns the last category of tests, because it indicates that there are more efficient ways to handle those tests which the Java backend is taking advantage of but the Haskell backend is not. This is that list (note that we jump from 1.77x max slowdown in other tests to 2.82x slowdown in this list): ``` 2.82 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_29.json 2.87 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_11.json 2.89 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_8.json 2.90 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_1.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_28.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_9.json 2.94 tests/ethereum-tests/VMTests/vmArithmeticTest/fibbonacci_unrolled.json 2.95 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_23.json 2.96 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_0.json 2.97 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_16.json 2.98 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_25.json 2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_31.json 2.99 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_4.json 3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_21.json 3.01 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_32.json 3.02 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_2.json 3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_10.json 3.04 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_6.json 3.05 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_30.json 3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_12.json 3.06 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_18.json 3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_14.json 3.07 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_20.json 3.08 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_27.json 3.09 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_19.json 3.10 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_24.json 3.11 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_3.json 3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_17.json 3.12 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_22.json 3.13 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_33.json 3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_26.json 3.14 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_5.json 3.16 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_7.json 3.17 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_15.json 3.18 tests/ethereum-tests/VMTests/vmArithmeticTest/expPowerOf256Of256_13.json 3.69 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpdestBigList.json 4.86 tests/ethereum-tests/VMTests/vmIOandFlowOperations/byte1.json 5.37 tests/ethereum-tests/VMTests/vmIOandFlowOperations/jumpi_at_the_end.json ```
non_process
comparatively bad performance on some concrete tests when comparing the haskell backend to the java backend on kevm concrete tests tests total we have tests which the haskell backend finishes faster than the java backend tests which finish within the time of the java backend slowdown tests which timeout unexpectedly see and tests which take more than the time of the java backend this issue concerns the last category of tests because it indicates that there are more efficient ways to handle those tests which the java backend is taking advantage of but the haskell backend is not this is that list note that we jump from max slowdown in other tests to slowdown in this list tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest fibbonacci unrolled json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmarithmetictest json tests ethereum tests vmtests vmioandflowoperations jumpdestbiglist json tests ethereum tests vmtests vmioandflowoperations json tests ethereum tests vmtests vmioandflowoperations jumpi at the end json
0
145,816
11,709,059,966
IssuesEvent
2020-03-08 16:43:44
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Reflect scan state with the right color in the UI
[zube]: To Test area/scan-tool team/ui
UI counterpart for https://github.com/rancher/rancher/issues/25214 UI needs to show "Fail" in Red and "Pass" in Green
1.0
Reflect scan state with the right color in the UI - UI counterpart for https://github.com/rancher/rancher/issues/25214 UI needs to show "Fail" in Red and "Pass" in Green
non_process
reflect scan state with the right color in the ui ui counterpart for ui needs to show fail in red and pass in green
0
20,482
27,140,584,979
IssuesEvent
2023-02-16 16:04:44
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Add ignoreList to redaction processor
enhancement processor/redaction needs triage
### Component(s) processor/redaction ### Is your feature request related to a problem? Please describe. Some fields may have a 12 digit number but it's an internal system identifier which doesn't benefit from redaction. It would be nice if these attributes could be passed through without having the value treated as a credit card number. ### Describe the solution you'd like Add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data. ### Describe alternatives you've considered Other processors don't seem to have the regex matching and replacement ability that the redaction processor has. ### Additional context The "allowList" in the processor seems to provide this capabillity at first glance but it is actually different. Adding the ignoreList will give users a way to proceed without having the modify their system identifiers or other drastic measures.
1.0
Add ignoreList to redaction processor - ### Component(s) processor/redaction ### Is your feature request related to a problem? Please describe. Some fields may have a 12 digit number but it's an internal system identifier which doesn't benefit from redaction. It would be nice if these attributes could be passed through without having the value treated as a credit card number. ### Describe the solution you'd like Add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data. ### Describe alternatives you've considered Other processors don't seem to have the regex matching and replacement ability that the redaction processor has. ### Additional context The "allowList" in the processor seems to provide this capabillity at first glance but it is actually different. Adding the ignoreList will give users a way to proceed without having the modify their system identifiers or other drastic measures.
process
add ignorelist to redaction processor component s processor redaction is your feature request related to a problem please describe some fields may have a digit number but it s an internal system identifier which doesn t benefit from redaction it would be nice if these attributes could be passed through without having the value treated as a credit card number describe the solution you d like add an ignorelist to the redaction processor so it can be explicitly set to pass suspicious data describe alternatives you ve considered other processors don t seem to have the regex matching and replacement ability that the redaction processor has additional context the allowlist in the processor seems to provide this capabillity at first glance but it is actually different adding the ignorelist will give users a way to proceed without having the modify their system identifiers or other drastic measures
1
8,721
11,859,815,994
IssuesEvent
2020-03-25 13:58:55
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
opened
We should highlight that using prisma.raw() with parameters is not secure and recommend using prisma.raw``
kind/discussion kind/docs kind/improvement process/candidate topic: prisma-client
## Problem Users are using `prisma.raw()` like ```js const data = await prisma.raw( `SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;`, ); ``` This example is using `prisma.raw()` the pure text version so there is no security around parameters. Only raw`` is secure because it's using https://github.com/blakeembrey/sql-template-tag ## Solution In this case it would be recommended to do ```js const data = await prisma.raw` SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1; `; ``` This should be highlighted in the docs (and examples?) We also can think about how to warn users that are using `prisma.raw()` or even disable it under a flag? *Note prisma.raw`` parameters do not work as of today see https://github.com/prisma/prisma-client-js/issues/595*
1.0
We should highlight that using prisma.raw() with parameters is not secure and recommend using prisma.raw`` - ## Problem Users are using `prisma.raw()` like ```js const data = await prisma.raw( `SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1;`, ); ``` This example is using `prisma.raw()` the pure text version so there is no security around parameters. Only raw`` is secure because it's using https://github.com/blakeembrey/sql-template-tag ## Solution In this case it would be recommended to do ```js const data = await prisma.raw` SELECT * FROM "ProviderItemAttribute" WHERE "provider_item" = ${root.id} AND "user" = ${auth.user.id} limit 1; `; ``` This should be highlighted in the docs (and examples?) We also can think about how to warn users that are using `prisma.raw()` or even disable it under a flag? *Note prisma.raw`` parameters do not work as of today see https://github.com/prisma/prisma-client-js/issues/595*
process
we should highlight that using prisma raw with parameters is not secure and recommend using prisma raw problem users are using prisma raw like js const data await prisma raw select from provideritemattribute where provider item root id and user auth user id limit this example is using prisma raw the pure text version so there is no security around parameters only raw is secure because it s using solution in this case it would be recommended to do js const data await prisma raw select from provideritemattribute where provider item root id and user auth user id limit this should be highlighted in the docs and examples we also can think about how to warn users that are using prisma raw or even disable it under a flag note prisma raw parameters do not work as of today see
1
86,705
10,788,967,321
IssuesEvent
2019-11-05 10:51:30
buildsofwarcraft/app
https://api.github.com/repos/buildsofwarcraft/app
opened
Approval from Overwolf
design overwolf
- [ ] ask for approval from Overwolf regarding first design draft - [ ] implement feedback from Overwolf if necessary
1.0
Approval from Overwolf - - [ ] ask for approval from Overwolf regarding first design draft - [ ] implement feedback from Overwolf if necessary
non_process
approval from overwolf ask for approval from overwolf regarding first design draft implement feedback from overwolf if necessary
0
19,825
26,214,749,921
IssuesEvent
2023-01-04 10:01:35
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Add relations export option to package alg (Request in QGIS)
Easy fix Processing Alg 3.28
### Request for documentation From pull request QGIS/qgis#49520 Author: @elpaso QGIS version: 3.28 **Add relations export option to package alg** ### PR Description: Adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream, taking selected features into account. ![immagine](https://user-images.githubusercontent.com/142164/182017853-9d4eac26-be78-4910-986a-7ecd991478bf.png) Funded by: ARPA Piemonte ### Commits tagged with [need-docs] or [FEATURE]
1.0
Add relations export option to package alg (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#49520 Author: @elpaso QGIS version: 3.28 **Add relations export option to package alg** ### PR Description: Adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream, taking selected features into account. ![immagine](https://user-images.githubusercontent.com/142164/182017853-9d4eac26-be78-4910-986a-7ecd991478bf.png) Funded by: ARPA Piemonte ### Commits tagged with [need-docs] or [FEATURE]
process
add relations export option to package alg request in qgis request for documentation from pull request qgis qgis author elpaso qgis version add relations export option to package alg pr description adds an option to the package algorithm to follow project relations and export related tables following the hierarchy upstream and downstream taking selected features into account funded by arpa piemonte commits tagged with or
1
19,626
10,475,840,969
IssuesEvent
2019-09-23 17:14:03
chapel-lang/chapel
https://api.github.com/repos/chapel-lang/chapel
closed
Slow Sparse Matrix Addition
area: Libraries / Modules type: Performance user issue
I tried using `.plus()` to add a large (square 10^6 with 10^6 nonzeros) matrix to it's transpose using `var ppt = P.plus(PT);` and the computation hadn't finished after 45 minutes so I killed it. I wrote my own matrix addition method to work around this as follows ```` proc matPlus(A:[],B:[]) { var dom = {A.domain.dim(1),B.domain.dim(2)}; var sps = CSRDomain(dom); sps += A.domain; sps += B.domain; var S: [sps] real; for (i,j) in sps { S(i,j) = A(i,j) + B(i,j); } return S; } ```` and with that instead, the same line ran in 26 seconds.
True
Slow Sparse Matrix Addition - I tried using `.plus()` to add a large (square 10^6 with 10^6 nonzeros) matrix to it's transpose using `var ppt = P.plus(PT);` and the computation hadn't finished after 45 minutes so I killed it. I wrote my own matrix addition method to work around this as follows ```` proc matPlus(A:[],B:[]) { var dom = {A.domain.dim(1),B.domain.dim(2)}; var sps = CSRDomain(dom); sps += A.domain; sps += B.domain; var S: [sps] real; for (i,j) in sps { S(i,j) = A(i,j) + B(i,j); } return S; } ```` and with that instead, the same line ran in 26 seconds.
non_process
slow sparse matrix addition i tried using plus to add a large square with nonzeros matrix to it s transpose using var ppt p plus pt and the computation hadn t finished after minutes so i killed it i wrote my own matrix addition method to work around this as follows proc matplus a b var dom a domain dim b domain dim var sps csrdomain dom sps a domain sps b domain var s real for i j in sps s i j a i j b i j return s and with that instead the same line ran in seconds
0
155,927
5,962,703,358
IssuesEvent
2017-05-30 00:13:33
input-output-hk/cardano-sl
https://api.github.com/repos/input-output-hk/cardano-sl
closed
[CSL-148] Setup a Hydra instance as CI
Mirroring Priority:Normal State:Done Type:Task
<blockquote>@domen</blockquote> - Clean deploy (1TB EBS GP2, R3 instance) using nixops - Copy existing ssh keys - Slack integration - Declarative jobsets <!--MIRROR_META={"service":"youtrack","id":"CSL-148"}-->
1.0
[CSL-148] Setup a Hydra instance as CI - <blockquote>@domen</blockquote> - Clean deploy (1TB EBS GP2, R3 instance) using nixops - Copy existing ssh keys - Slack integration - Declarative jobsets <!--MIRROR_META={"service":"youtrack","id":"CSL-148"}-->
non_process
setup a hydra instance as ci domen clean deploy ebs instance using nixops copy existing ssh keys slack integration declarative jobsets
0
9,030
12,129,318,523
IssuesEvent
2020-04-22 22:17:41
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Kubernetes: Monitor API equivalent
enhancement process rest
**Problem** The monitor API tests the REST API by running queries against it and validating the results are within certain parameters. It then exposes those results via an API and a separate dashboard shows the results. The monitor has to be explicitly configured with the REST endpoints, which is not possible to know in a containerized environments where pods come and go (especially with HPA). It's also another component we'd have to containerize and maintain. Monitor tests: ``` Successfully called transactions with limit params only Successfully retrieved transactions from with 50 seconds ago Successfully retrieved balance from with 1800 seconds ago Successfully called transactions with time and limit params Successfully retrieved single transactions by id Successfully called balances with time and limit params Successfully called balances and performed account check Successfully called accounts with time and limit params Successfully called accounts for single account Successfully called accounts and performed account check Successfully called balances and performed account check Successfully called transactions with order params only ``` **Solution** - Add metrics that represent the tests that make sense - For sure, the `with X seconds ago` tests can be made as metrics in importer (might already have an equivalent) - Figure out if we should continue invoking every API ever X seconds or if a health check endpoint and helm acceptance test is enough for the others **Alternatives** - Containerize the monitor UI/API and have it list endpoints from k8s APIs **Additional Context**
1.0
Kubernetes: Monitor API equivalent - **Problem** The monitor API tests the REST API by running queries against it and validating the results are within certain parameters. It then exposes those results via an API and a separate dashboard shows the results. The monitor has to be explicitly configured with the REST endpoints, which is not possible to know in a containerized environments where pods come and go (especially with HPA). It's also another component we'd have to containerize and maintain. Monitor tests: ``` Successfully called transactions with limit params only Successfully retrieved transactions from with 50 seconds ago Successfully retrieved balance from with 1800 seconds ago Successfully called transactions with time and limit params Successfully retrieved single transactions by id Successfully called balances with time and limit params Successfully called balances and performed account check Successfully called accounts with time and limit params Successfully called accounts for single account Successfully called accounts and performed account check Successfully called balances and performed account check Successfully called transactions with order params only ``` **Solution** - Add metrics that represent the tests that make sense - For sure, the `with X seconds ago` tests can be made as metrics in importer (might already have an equivalent) - Figure out if we should continue invoking every API ever X seconds or if a health check endpoint and helm acceptance test is enough for the others **Alternatives** - Containerize the monitor UI/API and have it list endpoints from k8s APIs **Additional Context**
process
kubernetes monitor api equivalent problem the monitor api tests the rest api by running queries against it and validating the results are within certain parameters it then exposes those results via an api and a separate dashboard shows the results the monitor has to be explicitly configured with the rest endpoints which is not possible to know in a containerized environments where pods come and go especially with hpa it s also another component we d have to containerize and maintain monitor tests successfully called transactions with limit params only successfully retrieved transactions from with seconds ago successfully retrieved balance from with seconds ago successfully called transactions with time and limit params successfully retrieved single transactions by id successfully called balances with time and limit params successfully called balances and performed account check successfully called accounts with time and limit params successfully called accounts for single account successfully called accounts and performed account check successfully called balances and performed account check successfully called transactions with order params only solution add metrics that represent the tests that make sense for sure the with x seconds ago tests can be made as metrics in importer might already have an equivalent figure out if we should continue invoking every api ever x seconds or if a health check endpoint and helm acceptance test is enough for the others alternatives containerize the monitor ui api and have it list endpoints from apis additional context
1
223,855
24,752,782,168
IssuesEvent
2022-10-21 14:59:43
turkdevops/bit
https://api.github.com/repos/turkdevops/bit
closed
CVE-2015-9251 (Low) detected in jquery-1.7.2.min.js, jquery-1.9.0.min.js - autoclosed
security vulnerability
## CVE-2015-9251 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.9.0.min.js</b></p></summary> <p> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.9.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/chai-string/test/index.html</p> <p>Path to vulnerable library: /node_modules/chai-string/test/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.0.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/bit/commit/2d5ba447ad6b7168a8a152ec95ebf4c142d6441e">2d5ba447ad6b7168a8a152ec95ebf4c142d6441e</a></p> <p>Found in base branch: <b>app-generator</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Low) detected in jquery-1.7.2.min.js, jquery-1.9.0.min.js - autoclosed - ## CVE-2015-9251 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.9.0.min.js</b></p></summary> <p> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.9.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/chai-string/test/index.html</p> <p>Path to vulnerable library: /node_modules/chai-string/test/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.0.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/bit/commit/2d5ba447ad6b7168a8a152ec95ebf4c142d6441e">2d5ba447ad6b7168a8a152ec95ebf4c142d6441e</a></p> <p>Found in base branch: <b>app-generator</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve low detected in jquery min js jquery min js autoclosed cve low severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file node modules jmespath index html path to vulnerable library node modules jmespath index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node modules chai string test index html path to vulnerable library node modules chai string test index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch app generator vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
0
27,585
11,516,145,089
IssuesEvent
2020-02-14 03:51:41
MadeByEmil/wdhan-basic
https://api.github.com/repos/MadeByEmil/wdhan-basic
closed
CVE-2015-9251 (Medium) detected in jquery-1.10.2.min.js
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/wdhan-basic/_includes/head.html</p> <p>Path to vulnerable library: /wdhan-basic/_includes/head.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.10.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MadeByEmil/wdhan-basic/commit/20319c55082efc29a29a8cb4e602f2bb13bcd75f">20319c55082efc29a29a8cb4e602f2bb13bcd75f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.10.2.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/wdhan-basic/_includes/head.html</p> <p>Path to vulnerable library: /wdhan-basic/_includes/head.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.10.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MadeByEmil/wdhan-basic/commit/20319c55082efc29a29a8cb4e602f2bb13bcd75f">20319c55082efc29a29a8cb4e602f2bb13bcd75f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm wdhan basic includes head html path to vulnerable library wdhan basic includes head html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
89,070
11,194,813,953
IssuesEvent
2020-01-03 03:04:46
ChildMindInstitute/mindlogger-app
https://api.github.com/repos/ChildMindInstitute/mindlogger-app
closed
ML-573 ⁃ MindLogger log in asks for username and email
Android Eden2016 design frontend iOS
Only username works to log in, but box prompts for email ![IMG_4295](https://user-images.githubusercontent.com/48026772/70557857-ab644e80-1b51-11ea-8374-a13fe4c96711.PNG)
1.0
ML-573 ⁃ MindLogger log in asks for username and email - Only username works to log in, but box prompts for email ![IMG_4295](https://user-images.githubusercontent.com/48026772/70557857-ab644e80-1b51-11ea-8374-a13fe4c96711.PNG)
non_process
ml ⁃ mindlogger log in asks for username and email only username works to log in but box prompts for email
0
131,177
18,214,827,622
IssuesEvent
2021-09-30 02:01:16
samjcs/xcp-ng-org
https://api.github.com/repos/samjcs/xcp-ng-org
opened
CVE-2018-20676 (Medium) detected in bootstrap-3.3.5.min.js
security vulnerability
## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p> <p>Path to dependency file: xcp-ng-org/node_modules/autocomplete.js/test/playground_jquery.html</p> <p>Path to vulnerable library: /node_modules/autocomplete.js/test/playground_jquery.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samjcs/xcp-ng-org/commit/76bae0317ed38f295fb61df7a9af82f659a38c78">76bae0317ed38f295fb61df7a9af82f659a38c78</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.5","packageFilePaths":["/node_modules/autocomplete.js/test/playground_jquery.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-20676 (Medium) detected in bootstrap-3.3.5.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p> <p>Path to dependency file: xcp-ng-org/node_modules/autocomplete.js/test/playground_jquery.html</p> <p>Path to vulnerable library: /node_modules/autocomplete.js/test/playground_jquery.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samjcs/xcp-ng-org/commit/76bae0317ed38f295fb61df7a9af82f659a38c78">76bae0317ed38f295fb61df7a9af82f659a38c78</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.5","packageFilePaths":["/node_modules/autocomplete.js/test/playground_jquery.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file xcp ng org node modules autocomplete js test playground jquery html path to vulnerable library node modules autocomplete js test playground jquery html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree twitter bootstrap isminimumfixversionavailable true minimumfixversion bootstrap basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the tooltip data viewport attribute vulnerabilityurl
0
113,845
24,499,057,831
IssuesEvent
2022-10-10 11:14:14
assemblee-virtuelle/semapps
https://api.github.com/repos/assemblee-virtuelle/semapps
closed
Partager comment remonter des erreurs sur Sentry
5. Qualité de code 11. Infoculteur low priority
https://sentry.io/welcome/ Fait par @srosset81 sur le serveur Colibris, à partager avec les autres. Eventuellement le mettre en place sur Archipel ?
1.0
Partager comment remonter des erreurs sur Sentry - https://sentry.io/welcome/ Fait par @srosset81 sur le serveur Colibris, à partager avec les autres. Eventuellement le mettre en place sur Archipel ?
non_process
partager comment remonter des erreurs sur sentry fait par sur le serveur colibris à partager avec les autres eventuellement le mettre en place sur archipel
0
2,823
5,773,309,470
IssuesEvent
2017-04-28 01:27:06
gaocegege/maintainer
https://api.github.com/repos/gaocegege/maintainer
opened
Detect the language for CONTRIBUTING
process/not claimed type/feature
It is the base of language specific CONTRIBUTING generation.
1.0
Detect the language for CONTRIBUTING - It is the base of language specific CONTRIBUTING generation.
process
detect the language for contributing it is the base of language specific contributing generation
1
4,580
7,410,057,882
IssuesEvent
2018-03-21 00:01:01
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Video quality only low -> unreadeable
active-directory assigned-to-author docs-experience in-process triaged
Hi, Not sure what's wrong, but the videoplayer informs me only "Low" quality is avaialble for the videos, so the screens are not readeable. Valentijn --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 751889da-f001-117f-d525-3b27d467edf2 * Version Independent ID: ce46adec-de86-7c34-45ba-d735cba76d2d * Content: [How do Azure Active Directory admins add B2B collaboration users?](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-b2b-admin-add-users#feedback) * Content Source: [articles/active-directory/active-directory-b2b-admin-add-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-b2b-admin-add-users.md) * Service: **active-directory** * GitHub Login: @twooley * Microsoft Alias: **twooley**
1.0
Video quality only low -> unreadeable - Hi, Not sure what's wrong, but the videoplayer informs me only "Low" quality is avaialble for the videos, so the screens are not readeable. Valentijn --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 751889da-f001-117f-d525-3b27d467edf2 * Version Independent ID: ce46adec-de86-7c34-45ba-d735cba76d2d * Content: [How do Azure Active Directory admins add B2B collaboration users?](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-b2b-admin-add-users#feedback) * Content Source: [articles/active-directory/active-directory-b2b-admin-add-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-b2b-admin-add-users.md) * Service: **active-directory** * GitHub Login: @twooley * Microsoft Alias: **twooley**
process
video quality only low unreadeable hi not sure what s wrong but the videoplayer informs me only low quality is avaialble for the videos so the screens are not readeable valentijn document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service active directory github login twooley microsoft alias twooley
1
323,174
23,937,399,790
IssuesEvent
2022-09-11 12:33:25
os-climate/ITR
https://api.github.com/repos/os-climate/ITR
opened
Units: mboe vs mmboe
documentation
We all know that "boe" is a "Barrel of Oil Equivalent". It appears that the 'm' in "mboe" is a Roman M, meaning thousand, and that 'mm' is a Roman thousand squared, meaning a million. Is that correct? Is 'mcf' therefore a thousand cu ft of gas? Are there other 'm's that are thousands instead of millions we need to look out for? I will need to update our units files (which don't know this particular prefixing rule).
1.0
Units: mboe vs mmboe - We all know that "boe" is a "Barrel of Oil Equivalent". It appears that the 'm' in "mboe" is a Roman M, meaning thousand, and that 'mm' is a Roman thousand squared, meaning a million. Is that correct? Is 'mcf' therefore a thousand cu ft of gas? Are there other 'm's that are thousands instead of millions we need to look out for? I will need to update our units files (which don't know this particular prefixing rule).
non_process
units mboe vs mmboe we all know that boe is a barrel of oil equivalent it appears that the m in mboe is a roman m meaning thousand and that mm is a roman thousand squared meaning a million is that correct is mcf therefore a thousand cu ft of gas are there other m s that are thousands instead of millions we need to look out for i will need to update our units files which don t know this particular prefixing rule
0
192,269
15,342,345,966
IssuesEvent
2021-02-27 15:51:44
jsinger0420/NodeEra
https://api.github.com/repos/jsinger0420/NodeEra
opened
Python / QT development environment
documentation
Need to write instructions on how to install and run the application from python.
1.0
Python / QT development environment - Need to write instructions on how to install and run the application from python.
non_process
python qt development environment need to write instructions on how to install and run the application from python
0
15,730
10,265,612,819
IssuesEvent
2019-08-22 19:17:41
ualbertalib/avalon
https://api.github.com/repos/ualbertalib/avalon
closed
Move Publish/unpublish, Edit and Delete buttons to the right side of objects and make more prominent
Post-launch usability
### Descriptive summary It is easy for users to miss where the publish, edit and delete buttons are ### Expected behavior Publish, edit and delete buttons are right justified on object views Buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme ### Actual behavior Publish, edit and delete buttons are left justified and easy to miss ### Steps to reproduce the behavior 1. Go to any record view and observe the placement and style of these buttons
True
Move Publish/unpublish, Edit and Delete buttons to the right side of objects and make more prominent - ### Descriptive summary It is easy for users to miss where the publish, edit and delete buttons are ### Expected behavior Publish, edit and delete buttons are right justified on object views Buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme ### Actual behavior Publish, edit and delete buttons are left justified and easy to miss ### Steps to reproduce the behavior 1. Go to any record view and observe the placement and style of these buttons
non_process
move publish unpublish edit and delete buttons to the right side of objects and make more prominent descriptive summary it is easy for users to miss where the publish edit and delete buttons are expected behavior publish edit and delete buttons are right justified on object views buttons are easier to see with better styling such as thicker borders and more contrasting colour scheme actual behavior publish edit and delete buttons are left justified and easy to miss steps to reproduce the behavior go to any record view and observe the placement and style of these buttons
0
5,344
8,176,038,838
IssuesEvent
2018-08-28 05:48:39
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
closed
Update workflow/process documents
a-Process c.DevOps
There has been some internal restructuring as well as improvements to GitHub that necessitates updates to the workflow/process document. To name a few: - No longer need to ask dev to update branch if outdated since GitHub provides ["Allow edits from maintainers"](https://github.com/blog/2247-improving-collaboration-with-forks) (since very long ago) - [Multiple issue/PR templates](https://github.com/blog/2495-multiple-issue-and-pull-request-templates) - [Label descriptions](https://github.com/blog/2505-label-improvements-emoji-descriptions-and-more) (already adopted but the docs need update) - Removing local test - Removing new contributor intro - Renaming `s.ToMerge` -> `s.FinalReview` and `s.MergeApproved` -> `s.ToMerge` Also, the core team structure has changed. Will take some time this weekend to update.
1.0
Update workflow/process documents - There has been some internal restructuring as well as improvements to GitHub that necessitates updates to the workflow/process document. To name a few: - No longer need to ask dev to update branch if outdated since GitHub provides ["Allow edits from maintainers"](https://github.com/blog/2247-improving-collaboration-with-forks) (since very long ago) - [Multiple issue/PR templates](https://github.com/blog/2495-multiple-issue-and-pull-request-templates) - [Label descriptions](https://github.com/blog/2505-label-improvements-emoji-descriptions-and-more) (already adopted but the docs need update) - Removing local test - Removing new contributor intro - Renaming `s.ToMerge` -> `s.FinalReview` and `s.MergeApproved` -> `s.ToMerge` Also, the core team structure has changed. Will take some time this weekend to update.
process
update workflow process documents there has been some internal restructuring as well as improvements to github that necessitates updates to the workflow process document to name a few no longer need to ask dev to update branch if outdated since github provides since very long ago already adopted but the docs need update removing local test removing new contributor intro renaming s tomerge s finalreview and s mergeapproved s tomerge also the core team structure has changed will take some time this weekend to update
1
7,907
11,089,904,242
IssuesEvent
2019-12-14 22:01:00
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
topicpull stylesheet error
bug preprocess preprocess2 stale
Hi, I got the stylesheet error when running my test data. ``` topicpull: [topicpull] Transforming into D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp [topicpull] Loading stylesheet D:\DITA-OT\dita-ot-2.4.6\xsl\preprocess\topicpull.xsl [topicpull] Processing D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp\topics-en\p_programmingelements.xml [topicpull] Error at xsl:param on line 623 of topicpullImpl.xsl: [topicpull] XTDE0610: A value must be supplied for the parameter because the default value is not a [topicpull] valid instance of the required type [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#96) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#55) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1] [topicpull] in built-in template rule [topicpull] Failed to transform document: A value must be supplied for the parameter because the default value is not a valid instance of the required type ``` The test data is here: [20170419-sample-en-for-word.zip](https://github.com/dita-ot/dita-ot/files/1017504/20170419-sample-en-for-word.zip) The command-line is here: ``` PS D:\DITA-OT\dita-ot-2.4.6> bin\dita -i "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\sample_en.ditamap" -f pdf2 -o "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\out\pdf2" -l "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\log.txt" -v -t "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp" --pdf.formatter=ah Unable to locate tools.jar. Expected to find it in C:\Program Files\Java\jre1.8.0_131\lib\tools.jar PS D:\DITA-OT\dita-ot-2.4.6> ``` As this data was written in the past, If it contains authoring mistake, please let me know: Regards, -- /*-------------------------------------------------- Toshihiko Makita Development Group. Antenna House, Inc. Ina Branch Web site: http://www.antenna.co.jp/ http://www.antennahouse.com/ --------------------------------------------------*/
2.0
topicpull stylesheet error - Hi, I got the stylesheet error when running my test data. ``` topicpull: [topicpull] Transforming into D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp [topicpull] Loading stylesheet D:\DITA-OT\dita-ot-2.4.6\xsl\preprocess\topicpull.xsl [topicpull] Processing D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp\topics-en\p_programmingelements.xml [topicpull] Error at xsl:param on line 623 of topicpullImpl.xsl: [topicpull] XTDE0610: A value must be supplied for the parameter because the default value is not a [topicpull] valid instance of the required type [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#96) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpull-pr-d.xsl#55) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3]/fragref[1] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8]/syntaxdiagram[3] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1]/section[8] [topicpull] at xsl:apply-templates (file:/D:/DITA-OT/dita-ot-2.4.6/xsl/preprocess/topicpullImpl.xsl#1161) [topicpull] processing /topic/body[1] [topicpull] in built-in template rule [topicpull] Failed to transform document: A value must be supplied for the parameter because the default value is not a valid instance of the required type ``` The test data is here: [20170419-sample-en-for-word.zip](https://github.com/dita-ot/dita-ot/files/1017504/20170419-sample-en-for-word.zip) The command-line is here: ``` PS D:\DITA-OT\dita-ot-2.4.6> bin\dita -i "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\sample_en.ditamap" -f pdf2 -o "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\out\pdf2" -l "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\log.txt" -v -t "D:\SVN\pdf5\testdata\20170419-sample-en-for-word\temp" --pdf.formatter=ah Unable to locate tools.jar. Expected to find it in C:\Program Files\Java\jre1.8.0_131\lib\tools.jar PS D:\DITA-OT\dita-ot-2.4.6> ``` As this data was written in the past, If it contains authoring mistake, please let me know: Regards, -- /*-------------------------------------------------- Toshihiko Makita Development Group. Antenna House, Inc. Ina Branch Web site: http://www.antenna.co.jp/ http://www.antennahouse.com/ --------------------------------------------------*/
process
topicpull stylesheet error hi i got the stylesheet error when running my test data topicpull transforming into d svn testdata sample en for word temp loading stylesheet d dita ot dita ot xsl preprocess topicpull xsl processing d svn testdata sample en for word temp topics en p programmingelements xml error at xsl param on line of topicpullimpl xsl a value must be supplied for the parameter because the default value is not a valid instance of the required type at xsl apply templates file d dita ot dita ot xsl preprocess topicpull pr d xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpull pr d xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section syntaxdiagram fragref at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section syntaxdiagram at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body section at xsl apply templates file d dita ot dita ot xsl preprocess topicpullimpl xsl processing topic body in built in template rule failed to transform document a value must be supplied for the parameter because the default value is not a valid instance of the required type the test data is here the command line is here ps d dita ot dita ot bin dita i d svn testdata sample en for word sample en ditamap f o d svn testdata sample en for word out l d svn testdata sample en for word log txt v t d svn testdata sample en for word temp pdf formatter ah unable to locate tools jar expected to find it in c program files java lib tools jar ps d dita ot dita ot as this data was written in the past if it contains authoring mistake please let me know regards toshihiko makita development group antenna house inc ina branch web site
1
28,817
13,836,059,873
IssuesEvent
2020-10-14 00:07:12
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
CustomScrollView + SliverAppBar + WebView
P3 a: annoyance a: platform-views cp: 1.22 cp: 1.22 completed engine found in release: 1.20 has reproducible steps p: first party p: webview plugin severe: performance waiting for PR to land (fixed)
Hi, When trying to implement a floating `SliverAppBar` in conjunction with a `WebView` Flutter drops a significant amount of frames. It seems like the WebView is resizing itself whenever the `SliverAppBar` is scrolled out of view. ``` @override Widget build(BuildContext context) { return CustomScrollView( slivers: <Widget>[ SliverAppBar( title: const Text("Heading"), floating: true, ), SliverFillRemaining( child: WebView(initialUrl: "http://stackoverflow.com"), ) ], ); } ``` As seen on this GIF: ![flutter_1](https://user-images.githubusercontent.com/6739443/56343029-ecba7f00-61b9-11e9-963a-4e2bc3836f77.gif)
True
CustomScrollView + SliverAppBar + WebView - Hi, When trying to implement a floating `SliverAppBar` in conjunction with a `WebView` Flutter drops a significant amount of frames. It seems like the WebView is resizing itself whenever the `SliverAppBar` is scrolled out of view. ``` @override Widget build(BuildContext context) { return CustomScrollView( slivers: <Widget>[ SliverAppBar( title: const Text("Heading"), floating: true, ), SliverFillRemaining( child: WebView(initialUrl: "http://stackoverflow.com"), ) ], ); } ``` As seen on this GIF: ![flutter_1](https://user-images.githubusercontent.com/6739443/56343029-ecba7f00-61b9-11e9-963a-4e2bc3836f77.gif)
non_process
customscrollview sliverappbar webview hi when trying to implement a floating sliverappbar in conjunction with a webview flutter drops a significant amount of frames it seems like the webview is resizing itself whenever the sliverappbar is scrolled out of view override widget build buildcontext context return customscrollview slivers sliverappbar title const text heading floating true sliverfillremaining child webview initialurl as seen on this gif
0
11,959
14,726,198,674
IssuesEvent
2021-01-06 06:22:31
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Test failure: System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows
area-System.Diagnostics.Process
failed in job: [runtime-libraries-coreclr outerloop 20210105.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=938368&view=ms.vss-test-web.build-test-results-tab&runId=29665462&resultId=101960&paneView=debug) net6.0-windows-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Error message ~~~ System.UnauthorizedAccessException : Attempted to perform an unauthorized operation. Stack trace at System.Security.AccessControl.Win32.SetSecurityInfo(ResourceType type, String name, SafeHandle handle, SecurityInfos securityInformation, SecurityIdentifier owner, SecurityIdentifier group, GenericAcl sacl, GenericAcl dacl) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/Win32.cs:line 314 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, SafeHandle handle, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 263 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 353 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 343 at System.Security.AccessControl.FileSystemSecurity.Persist(String fullPath) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/Security/AccessControl/FileSystemSecurity.cs:line 124 at System.IO.FileSystemAclExtensions.SetAccessControl(FileInfo fileInfo, FileSecurity fileSecurity) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/IO/FileSystemAclExtensions.cs:line 78 at System.Diagnostics.Tests.ProcessStartInfoTests.SetAccessControl(String userName, String filePath, Boolean add) in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 534 at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 503 ~~~
1.0
Test failure: System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows - failed in job: [runtime-libraries-coreclr outerloop 20210105.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=938368&view=ms.vss-test-web.build-test-results-tab&runId=29665462&resultId=101960&paneView=debug) net6.0-windows-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Error message ~~~ System.UnauthorizedAccessException : Attempted to perform an unauthorized operation. Stack trace at System.Security.AccessControl.Win32.SetSecurityInfo(ResourceType type, String name, SafeHandle handle, SecurityInfos securityInformation, SecurityIdentifier owner, SecurityIdentifier group, GenericAcl sacl, GenericAcl dacl) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/Win32.cs:line 314 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, SafeHandle handle, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 263 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections, Object exceptionContext) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 353 at System.Security.AccessControl.NativeObjectSecurity.Persist(String name, AccessControlSections includeSections) in /_/src/libraries/System.Security.AccessControl/src/System/Security/AccessControl/NativeObjectSecurity.cs:line 343 at System.Security.AccessControl.FileSystemSecurity.Persist(String fullPath) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/Security/AccessControl/FileSystemSecurity.cs:line 124 at System.IO.FileSystemAclExtensions.SetAccessControl(FileInfo fileInfo, FileSecurity fileSecurity) in /_/src/libraries/System.IO.FileSystem.AccessControl/src/System/IO/FileSystemAclExtensions.cs:line 78 at System.Diagnostics.Tests.ProcessStartInfoTests.SetAccessControl(String userName, String filePath, Boolean add) in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 534 at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessStartInfoTests.cs:line 503 ~~~
process
test failure system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows failed in job windows release coreclr release windows server core open windows open mcr microsoft com dotnet buildtools prereqs windowsservercore helix error message system unauthorizedaccessexception attempted to perform an unauthorized operation stack trace at system security accesscontrol setsecurityinfo resourcetype type string name safehandle handle securityinfos securityinformation securityidentifier owner securityidentifier group genericacl sacl genericacl dacl in src libraries system security accesscontrol src system security accesscontrol cs line at system security accesscontrol nativeobjectsecurity persist string name safehandle handle accesscontrolsections includesections object exceptioncontext in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol nativeobjectsecurity persist string name accesscontrolsections includesections object exceptioncontext in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol nativeobjectsecurity persist string name accesscontrolsections includesections in src libraries system security accesscontrol src system security accesscontrol nativeobjectsecurity cs line at system security accesscontrol filesystemsecurity persist string fullpath in src libraries system io filesystem accesscontrol src system security accesscontrol filesystemsecurity cs line at system io filesystemaclextensions setaccesscontrol fileinfo fileinfo filesecurity filesecurity in src libraries system io filesystem accesscontrol src system io filesystemaclextensions cs line at system diagnostics tests processstartinfotests setaccesscontrol string username string filepath boolean add in src libraries system diagnostics process tests processstartinfotests cs line at system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows in src libraries system diagnostics process tests processstartinfotests cs line
1
7,998
11,188,267,117
IssuesEvent
2020-01-02 03:48:32
AlmuraDev/SGCraft
https://api.github.com/repos/AlmuraDev/SGCraft
closed
Stargates doesn't deal enough damage
bug in process
When I wear a Draconic Evolution armor the event horizon doesn't even hurt me. Could you increase it's damage to ensure that anything killable WILL die when they are at the wrong place at the wrong time?
1.0
Stargates doesn't deal enough damage - When I wear a Draconic Evolution armor the event horizon doesn't even hurt me. Could you increase it's damage to ensure that anything killable WILL die when they are at the wrong place at the wrong time?
process
stargates doesn t deal enough damage when i wear a draconic evolution armor the event horizon doesn t even hurt me could you increase it s damage to ensure that anything killable will die when they are at the wrong place at the wrong time
1
35,399
14,681,673,078
IssuesEvent
2020-12-31 14:00:24
eventespresso/barista
https://api.github.com/repos/eventespresso/barista
closed
Investigate isRTL from i18n and check Next component
C: services 🤝 D: Packages 📦 P3: med priority 😐 T: bug 🐞 T: task 🧹
Currently `isRTL()` is returning `false` when a RTL language is used in WP. The same result is obtained with `wp.i18n.isRTL()`
1.0
Investigate isRTL from i18n and check Next component - Currently `isRTL()` is returning `false` when a RTL language is used in WP. The same result is obtained with `wp.i18n.isRTL()`
non_process
investigate isrtl from and check next component currently isrtl is returning false when a rtl language is used in wp the same result is obtained with wp isrtl
0
2,652
5,429,403,363
IssuesEvent
2017-03-03 18:22:08
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Missing or truncated error message
confirmed-bug process
This looks close to #6456, and I thought it was just another case of that issue, but @indutny mentioned that the errors are printed from c++, so #6456 alone shouldn't have caused this, this is why I'm opening a separate issue. A harder to reproduce example where the error is missing (it reproduces with about 10% chance for me): ``` js for (var i = 0; i < 10000; i++) { console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i); } a(); ``` A simplier to reproduce example (the error gets truncated here most of the times, and sometimes is missing as in the previous example): ``` js for (var i = 0; i < 10000; i++) { console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i); } throw new Error(Array(100000 + 1).join('x') + '!'); ```
1.0
Missing or truncated error message - This looks close to #6456, and I thought it was just another case of that issue, but @indutny mentioned that the errors are printed from c++, so #6456 alone shouldn't have caused this, this is why I'm opening a separate issue. A harder to reproduce example where the error is missing (it reproduces with about 10% chance for me): ``` js for (var i = 0; i < 10000; i++) { console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i); } a(); ``` A simplier to reproduce example (the error gets truncated here most of the times, and sometimes is missing as in the previous example): ``` js for (var i = 0; i < 10000; i++) { console.log('HelloHelloHelloHelloHelloHelelloHelloHelloHelloHelloHelloHello ' + i); } throw new Error(Array(100000 + 1).join('x') + '!'); ```
process
missing or truncated error message this looks close to and i thought it was just another case of that issue but indutny mentioned that the errors are printed from c so alone shouldn t have caused this this is why i m opening a separate issue a harder to reproduce example where the error is missing it reproduces with about chance for me js for var i i i console log hellohellohellohellohellohelellohellohellohellohellohellohello i a a simplier to reproduce example the error gets truncated here most of the times and sometimes is missing as in the previous example js for var i i i console log hellohellohellohellohellohelellohellohellohellohellohellohello i throw new error array join x
1
59,224
14,369,086,827
IssuesEvent
2020-12-01 09:18:45
ignatandrei/stankins
https://api.github.com/repos/ignatandrei/stankins
closed
CVE-2019-6284 (Medium) detected in multiple libraries
security vulnerability
## CVE-2019-6284 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.10.0.tgz</b>, <b>node-sass-4.9.3.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.10.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.11.4.tgz (Root Library) - :x: **node-sass-4.10.0.tgz** (Vulnerable Library) </details> <details><summary><b>node-sass-4.9.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.10.7.tgz (Root Library) - :x: **node-sass-4.9.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-6284 (Medium) detected in multiple libraries - ## CVE-2019-6284 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.10.0.tgz</b>, <b>node-sass-4.9.3.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.10.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.10.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.11.4.tgz (Root Library) - :x: **node-sass-4.10.0.tgz** (Vulnerable Library) </details> <details><summary><b>node-sass-4.9.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.10.7.tgz (Root Library) - :x: **node-sass-4.9.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm stankins solution stankinsdatawebangular package json path to vulnerable library tmp ws scm stankins solution stankinsdatawebangular node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm stankins solution stankinsaliveangular package json path to vulnerable library tmp ws scm stankins solution stankinsaliveangular node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a heap based buffer over read exists in sass prelexer alternatives in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
7,527
10,600,052,478
IssuesEvent
2019-10-10 09:16:25
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Analyse and define how app frontend and app backend should handle process
analysis api-expose app-backend app-frontend process
As part of #243 there is defined some new API's for handling of process. There are some decisions that needs to be made on how the process api should work and how App Frontend should use the API The below diagram is a suggestion and are used as input for defining the clarification needed ![image](https://user-images.githubusercontent.com/13309071/64004678-ff04dc00-cb0e-11e9-8648-bbdfcd6ed7d8.png) The flow above is 1. User clicks link to a App from service catalog 2. The browser redirect to the start service page in the app. The react application renders the correct view 3. User click on instantiate button. 4. App Frontend calls instansiate API 5. Instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to. 6. User/browser open the instance 7. App frontend calls app backed to get the instance state 8. App Frontend render the correct component for the given state. 9. User input data . App Frontend call Data API to update data. If calculation or dataretrieavel causes additional data updates on server side, the response indicates to app frontend to get the latest data 10. App frontend gets the updated data 9 and 10 is repeated as long user has data to input 11. User press complete/send inn button 12. App frontend Saves data 13. App frontend calls validation API and any possible errors are presented to the user 14. If no errors are returned App Frontend call Process API to complete Task. Process Api trigger task validation to make sure it is allowed to move process forward. If validation fails, process API returns ### Define interaction between API and consumers (App Frontend, end user systems, agency systems) We need to define exactly how App Frontend and End User System will interact with process ### App logic for task completion We need to define what kind of control the app developer should Should we let the App developer add code to a HandleTaskCompletion ### Error message when task in process is tried to be completed When App Frontend post to Process API to complete task it might be condition for the current task that is not valid. It could be validation errors. Should validation errors be return or not?
1.0
Analyse and define how app frontend and app backend should handle process - As part of #243 there is defined some new API's for handling of process. There are some decisions that needs to be made on how the process api should work and how App Frontend should use the API The below diagram is a suggestion and are used as input for defining the clarification needed ![image](https://user-images.githubusercontent.com/13309071/64004678-ff04dc00-cb0e-11e9-8648-bbdfcd6ed7d8.png) The flow above is 1. User clicks link to a App from service catalog 2. The browser redirect to the start service page in the app. The react application renders the correct view 3. User click on instantiate button. 4. App Frontend calls instansiate API 5. Instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to. 6. User/browser open the instance 7. App frontend calls app backed to get the instance state 8. App Frontend render the correct component for the given state. 9. User input data . App Frontend call Data API to update data. If calculation or dataretrieavel causes additional data updates on server side, the response indicates to app frontend to get the latest data 10. App frontend gets the updated data 9 and 10 is repeated as long user has data to input 11. User press complete/send inn button 12. App frontend Saves data 13. App frontend calls validation API and any possible errors are presented to the user 14. If no errors are returned App Frontend call Process API to complete Task. Process Api trigger task validation to make sure it is allowed to move process forward. If validation fails, process API returns ### Define interaction between API and consumers (App Frontend, end user systems, agency systems) We need to define exactly how App Frontend and End User System will interact with process ### App logic for task completion We need to define what kind of control the app developer should Should we let the App developer add code to a HandleTaskCompletion ### Error message when task in process is tried to be completed When App Frontend post to Process API to complete task it might be condition for the current task that is not valid. It could be validation errors. Should validation errors be return or not?
process
analyse and define how app frontend and app backend should handle process as part of there is defined some new api s for handling of process there are some decisions that needs to be made on how the process api should work and how app frontend should use the api the below diagram is a suggestion and are used as input for defining the clarification needed the flow above is user clicks link to a app from service catalog the browser redirect to the start service page in the app the react application renders the correct view user click on instantiate button app frontend calls instansiate api instantiate create a new instance with data for the datamodell with prefill and returns instance information that user should be redirect to user browser open the instance app frontend calls app backed to get the instance state app frontend render the correct component for the given state user input data app frontend call data api to update data if calculation or dataretrieavel causes additional data updates on server side the response indicates to app frontend to get the latest data app frontend gets the updated data and is repeated as long user has data to input user press complete send inn button app frontend saves data app frontend calls validation api and any possible errors are presented to the user if no errors are returned app frontend call process api to complete task process api trigger task validation to make sure it is allowed to move process forward if validation fails process api returns define interaction between api and consumers app frontend end user systems agency systems we need to define exactly how app frontend and end user system will interact with process app logic for task completion we need to define what kind of control the app developer should should we let the app developer add code to a handletaskcompletion error message when task in process is tried to be completed when app frontend post to process api to complete task it might be condition for the current task that is not valid it could be validation errors should validation errors be return or not
1
6,222
9,160,890,597
IssuesEvent
2019-03-01 08:58:36
fablabbcn/fablabs.io
https://api.github.com/repos/fablabbcn/fablabs.io
closed
Bug in the labs list when adding a new lab
Approval Process bug
When adding a new lab, in the labs list there are some strange items (more than one!): ![screen shot 2016-12-14 at 15 34 28](https://cloud.githubusercontent.com/assets/1458229/21192825/220dd382-c22a-11e6-936e-0d3767a21d77.png)
1.0
Bug in the labs list when adding a new lab - When adding a new lab, in the labs list there are some strange items (more than one!): ![screen shot 2016-12-14 at 15 34 28](https://cloud.githubusercontent.com/assets/1458229/21192825/220dd382-c22a-11e6-936e-0d3767a21d77.png)
process
bug in the labs list when adding a new lab when adding a new lab in the labs list there are some strange items more than one
1
351,826
25,040,480,598
IssuesEvent
2022-11-04 20:10:15
CMPUT301F22T15/Foodverse
https://api.github.com/repos/CMPUT301F22T15/Foodverse
closed
Update CRC Cards For Part 3 Submission
documentation
- [x] Ensure CRC cards are up to date with classes in the code, updating where needed. - [x] #55
1.0
Update CRC Cards For Part 3 Submission - - [x] Ensure CRC cards are up to date with classes in the code, updating where needed. - [x] #55
non_process
update crc cards for part submission ensure crc cards are up to date with classes in the code updating where needed
0
11,965
14,729,400,203
IssuesEvent
2021-01-06 11:24:14
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Mobile App > Modify study list fields
Android P1 Process: Enhancement Process: Track 3 iOS
Modify study list to remove fields in accordance with #2600 . Adjust the UI. Have the study list display the default image provided by the Study Builder if there is no alternate image uploaded.
2.0
Mobile App > Modify study list fields - Modify study list to remove fields in accordance with #2600 . Adjust the UI. Have the study list display the default image provided by the Study Builder if there is no alternate image uploaded.
process
mobile app modify study list fields modify study list to remove fields in accordance with adjust the ui have the study list display the default image provided by the study builder if there is no alternate image uploaded
1
11,672
14,531,508,754
IssuesEvent
2020-12-14 20:54:05
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
"Import geotagged photos" fails for jpg / jpeg on Windows
Bug Processing Windows
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** `Import geotagged photos` algorithm fails for jpg / jpeg on Windows. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Unzip [sample_jpg.zip](https://github.com/qgis/QGIS/files/5688584/sample_jpg.zip) 2. Use the `Import geotagged photos` algorithm from the processing toolbox with the input folder containing the sample image. 4. See error --> `Could not open ...` **QGIS and OS versions** QGIS version | 3.16.1-Hannover | QGIS code revision | b381a90dca -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | Discovery; ImportPhotos-adapted_path; joinmultiplelines; mmqgis; QPackage; db_manager; MetaSearch; processing **Additional context** This happens for a lot of different images captured with different devices. On Linux it works with QGIS 3.16.1, but it's compiled against GDAL/OGR **2.4.0**. Might be related to GDAL but I'm not entirely sure (`gdalinfo` fails also: https://github.com/OSGeo/gdal/issues/3284). Also with several image viewers (e.g. GIMP) the sample jpg works flawlessly and it's possible to show the EXIF data.
1.0
"Import geotagged photos" fails for jpg / jpeg on Windows - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** `Import geotagged photos` algorithm fails for jpg / jpeg on Windows. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Unzip [sample_jpg.zip](https://github.com/qgis/QGIS/files/5688584/sample_jpg.zip) 2. Use the `Import geotagged photos` algorithm from the processing toolbox with the input folder containing the sample image. 4. See error --> `Could not open ...` **QGIS and OS versions** QGIS version | 3.16.1-Hannover | QGIS code revision | b381a90dca -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | Discovery; ImportPhotos-adapted_path; joinmultiplelines; mmqgis; QPackage; db_manager; MetaSearch; processing **Additional context** This happens for a lot of different images captured with different devices. On Linux it works with QGIS 3.16.1, but it's compiled against GDAL/OGR **2.4.0**. Might be related to GDAL but I'm not entirely sure (`gdalinfo` fails also: https://github.com/OSGeo/gdal/issues/3284). Also with several image viewers (e.g. GIMP) the sample jpg works flawlessly and it's possible to show the EXIF data.
process
import geotagged photos fails for jpg jpeg on windows bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug import geotagged photos algorithm fails for jpg jpeg on windows how to reproduce unzip use the import geotagged photos algorithm from the processing toolbox with the input folder containing the sample image see error could not open qgis and os versions qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins discovery importphotos adapted path joinmultiplelines mmqgis qpackage db manager metasearch processing additional context this happens for a lot of different images captured with different devices on linux it works with qgis but it s compiled against gdal ogr might be related to gdal but i m not entirely sure gdalinfo fails also also with several image viewers e g gimp the sample jpg works flawlessly and it s possible to show the exif data
1
107,047
13,430,853,091
IssuesEvent
2020-09-07 05:51:14
ibmdb/go_ibm_db
https://api.github.com/repos/ibmdb/go_ibm_db
closed
New version does not work for parameterized queries
As Designed duplicate
We recently updated to new version 0.3.0, and existing queries are no longer working. Query like: `SELECT * FROM dummy_table WHERE cod_1 = ? AND cod_2 = ? AND cod_3 = '24'` Provides a **CLI0100E Wrong number of parameters. SQLSTATE=07001** I've been debugging a bit, and seen that provided args are not used in https://github.com/ibmdb/go_ibm_db/blob/6d3b2e8acd84b85fba599aa27246921a9b5482d2/conn.go#L50 Using DB2Driver 11.5 Any idea with that?
1.0
New version does not work for parameterized queries - We recently updated to new version 0.3.0, and existing queries are no longer working. Query like: `SELECT * FROM dummy_table WHERE cod_1 = ? AND cod_2 = ? AND cod_3 = '24'` Provides a **CLI0100E Wrong number of parameters. SQLSTATE=07001** I've been debugging a bit, and seen that provided args are not used in https://github.com/ibmdb/go_ibm_db/blob/6d3b2e8acd84b85fba599aa27246921a9b5482d2/conn.go#L50 Using DB2Driver 11.5 Any idea with that?
non_process
new version does not work for parameterized queries we recently updated to new version and existing queries are no longer working query like select from dummy table where cod and cod and cod provides a wrong number of parameters sqlstate i ve been debugging a bit and seen that provided args are not used in using any idea with that
0
15,767
19,913,145,692
IssuesEvent
2022-01-25 19:21:06
input-output-hk/high-assurance-legacy
https://api.github.com/repos/input-output-hk/high-assurance-legacy
closed
Automate the application of quasi-compatibility laws
type: enhancement reason: wontfix language: isabelle topic: process calculus
While in the ♮-calculus `▹` and `▹⇧∞` are compatible with bisimilarity, in the Þ-calculus they are not but only what we call quasi-compatible. While this still allows us to prove typical statements that we used to prove using compatibility, the proofs become quite verbose, since the lack of proper compatibility means that `process_family_equivalence` cannot rewrite under those combinators. We shall add a proof method that allows at least a limited form of rewriting under combinators like `▹` and `▹⇧∞` based on quasi-compatibility. Since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities, we shall name it `bisimilarity`. We shall keep `process_family_equivalence`, as `bisimilarity` will most likely not work with processes that perform branching based on received values.
1.0
Automate the application of quasi-compatibility laws - While in the ♮-calculus `▹` and `▹⇧∞` are compatible with bisimilarity, in the Þ-calculus they are not but only what we call quasi-compatible. While this still allows us to prove typical statements that we used to prove using compatibility, the proofs become quite verbose, since the lack of proper compatibility means that `process_family_equivalence` cannot rewrite under those combinators. We shall add a proof method that allows at least a limited form of rewriting under combinators like `▹` and `▹⇧∞` based on quasi-compatibility. Since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities, we shall name it `bisimilarity`. We shall keep `process_family_equivalence`, as `bisimilarity` will most likely not work with processes that perform branching based on received values.
process
automate the application of quasi compatibility laws while in the ♮ calculus ▹ and ▹⇧∞ are compatible with bisimilarity in the þ calculus they are not but only what we call quasi compatible while this still allows us to prove typical statements that we used to prove using compatibility the proofs become quite verbose since the lack of proper compatibility means that process family equivalence cannot rewrite under those combinators we shall add a proof method that allows at least a limited form of rewriting under combinators like ▹ and ▹⇧∞ based on quasi compatibility since this proof method will most likely use techniques that are specific to bisimilarity and thus limit the equivalences this proof method can work with to bisimilarities we shall name it bisimilarity we shall keep process family equivalence as bisimilarity will most likely not work with processes that perform branching based on received values
1
103,751
16,609,058,764
IssuesEvent
2021-06-02 09:14:50
dreamboy9/ofbiz
https://api.github.com/repos/dreamboy9/ofbiz
opened
CVE-2020-7656 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.3.2.min.js</b>, <b>jquery-1.8.1.min.js</b>, <b>jquery-1.6.2.js</b></p></summary> <p> <details><summary><b>jquery-1.3.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_standalone.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.3.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.6.2.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/index.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.6.2.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/ofbiz/commit/9f6ed39589395d00f1d69228cb50a7987ba11512">9f6ed39589395d00f1d69228cb50a7987ba11512</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7656 (Medium) detected in multiple libraries - ## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.3.2.min.js</b>, <b>jquery-1.8.1.min.js</b>, <b>jquery-1.6.2.js</b></p></summary> <p> <details><summary><b>jquery-1.3.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_standalone.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.3.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/tap/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.6.2.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.6.2/jquery.js</a></p> <p>Path to dependency file: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/index.html</p> <p>Path to vulnerable library: ofbiz/framework/images/webapp/images/jquery/plugins/validate/node_modules/qs/test/browser/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.6.2.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/ofbiz/commit/9f6ed39589395d00f1d69228cb50a7987ba11512">9f6ed39589395d00f1d69228cb50a7987ba11512</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery js jquery min js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules underscore string test test standalone html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules underscore string test test underscore vendor jquery js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules tap node modules redeyed examples browser index html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules tap node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file ofbiz framework images webapp images jquery plugins validate node modules qs test browser index html path to vulnerable library ofbiz framework images webapp images jquery plugins validate node modules qs test browser jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch trunk vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
68,485
21,664,953,483
IssuesEvent
2022-05-07 03:07:52
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Timestamp's position on Message Edits history modal window
T-Defect S-Tolerable A-Message-Editing A-Appearance O-Uncommon
### Steps to reproduce 1. Create a test room 2. Send a message 3. Edit the message 4. View the edit history ### Outcome #### What did you expect? The timestamp should be displayed on the left side. #### What happened instead? It is displayed on the center of the row. I guess the root cause is same as https://github.com/vector-im/element-web/issues/22011. ![before](https://user-images.githubusercontent.com/3362943/166224565-d19f71b5-b808-4f2d-8b89-3d8eca288f87.png) ### Operating system Debian ### Browser information Firefox ### URL for webapp localhost ### Application version 3e31fdb6a71f43774420e8da32452861296a263a ### Homeserver _No response_ ### Will you send logs? No
1.0
Timestamp's position on Message Edits history modal window - ### Steps to reproduce 1. Create a test room 2. Send a message 3. Edit the message 4. View the edit history ### Outcome #### What did you expect? The timestamp should be displayed on the left side. #### What happened instead? It is displayed on the center of the row. I guess the root cause is same as https://github.com/vector-im/element-web/issues/22011. ![before](https://user-images.githubusercontent.com/3362943/166224565-d19f71b5-b808-4f2d-8b89-3d8eca288f87.png) ### Operating system Debian ### Browser information Firefox ### URL for webapp localhost ### Application version 3e31fdb6a71f43774420e8da32452861296a263a ### Homeserver _No response_ ### Will you send logs? No
non_process
timestamp s position on message edits history modal window steps to reproduce create a test room send a message edit the message view the edit history outcome what did you expect the timestamp should be displayed on the left side what happened instead it is displayed on the center of the row i guess the root cause is same as operating system debian browser information firefox url for webapp localhost application version homeserver no response will you send logs no
0
185,751
6,727,384,932
IssuesEvent
2017-10-17 13:27:57
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
TinyTILE bluetooth app flash
area: Boards bug priority: low
**_Reported by Jie Zhou:_** Hello, I'm switching boards from arduino 101 to tinyTILE, the flashing method for tinyTILE is different than the arduino. Below are the commands I used for arduino flashing. ``` $ 1 make BOARD=tinytile $ 2 dfu-util -a x86_app -D outdir/arduino_101/zephyr.bin ``` I can make BOARD=tinyTILE. But when I enter in dfu-util -a x86_app -D outdir/tinytile/zephyr.bin the command line tells me outdir/tinytile/zephyr.bin cannot open file. Yet in outdir, after make BOARD command, tinytile directory is included. I would like to flash tinyTILE using dfu. Does anyone know the exact command? Thanks, Jie (Imported from Jira ZEP-2600)
1.0
TinyTILE bluetooth app flash - **_Reported by Jie Zhou:_** Hello, I'm switching boards from arduino 101 to tinyTILE, the flashing method for tinyTILE is different than the arduino. Below are the commands I used for arduino flashing. ``` $ 1 make BOARD=tinytile $ 2 dfu-util -a x86_app -D outdir/arduino_101/zephyr.bin ``` I can make BOARD=tinyTILE. But when I enter in dfu-util -a x86_app -D outdir/tinytile/zephyr.bin the command line tells me outdir/tinytile/zephyr.bin cannot open file. Yet in outdir, after make BOARD command, tinytile directory is included. I would like to flash tinyTILE using dfu. Does anyone know the exact command? Thanks, Jie (Imported from Jira ZEP-2600)
non_process
tinytile bluetooth app flash reported by jie zhou hello i m switching boards from arduino to tinytile the flashing method for tinytile is different than the arduino below are the commands i used for arduino flashing make board tinytile dfu util a app d outdir arduino zephyr bin i can make board tinytile but when i enter in dfu util a app d outdir tinytile zephyr bin the command line tells me outdir tinytile zephyr bin cannot open file yet in outdir after make board command tinytile directory is included i would like to flash tinytile using dfu does anyone know the exact command thanks jie imported from jira zep
0
46,053
5,996,412,311
IssuesEvent
2017-06-03 14:06:11
raulir/bccms
https://api.github.com/repos/raulir/bccms
closed
toolbar - put less used buttons under dropdown button
cms design/UX enhancement/refactoring
toolbar - put less used buttons under dropdown button
1.0
toolbar - put less used buttons under dropdown button - toolbar - put less used buttons under dropdown button
non_process
toolbar put less used buttons under dropdown button toolbar put less used buttons under dropdown button
0
20,934
27,779,912,770
IssuesEvent
2023-03-16 20:09:44
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
opened
Add functionality to profile, account settings, and login button on navbar
Processing Task Sprint 2
*Task Tests* run in "Sprint2-Navbar-Buttons" branch in github test1: - in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser - look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"![navbar-profile-account-login.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/5e880ed6-8151-4e1e-927d-96291bebd670) -click on the "profile" button to get to edit profile page:![edit-profile-img.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/8e5f5415-76d1-4978-b4ec-302f1e0c321d) -click on the "account settings" to get to the account settings page:![account-settings-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/a571095f-bd0c-4000-b3b8-5ed690986f0d) -click on "login" to get to login page:![login-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/ce877b30-5cea-4497-8a58-de3c627995b8)
1.0
Add functionality to profile, account settings, and login button on navbar - *Task Tests* run in "Sprint2-Navbar-Buttons" branch in github test1: - in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser - look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"![navbar-profile-account-login.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/5e880ed6-8151-4e1e-927d-96291bebd670) -click on the "profile" button to get to edit profile page:![edit-profile-img.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/8e5f5415-76d1-4978-b4ec-302f1e0c321d) -click on the "account settings" to get to the account settings page:![account-settings-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/a571095f-bd0c-4000-b3b8-5ed690986f0d) -click on "login" to get to login page:![login-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/ce877b30-5cea-4497-8a58-de3c627995b8)
process
add functionality to profile account settings and login button on navbar task tests run in navbar buttons branch in github in the project folder project cinco run npm start in the terminal this will open the homepage in a browser look at the navbar and ensure it has the following buttons profile account settings login click on the profile button to get to edit profile page click on the account settings to get to the account settings page click on login to get to login page
1
17,097
5,330,138,783
IssuesEvent
2017-02-15 16:24:38
eclipse/che
https://api.github.com/repos/eclipse/che
closed
The "Projects" info page does not refresh after deleting a project.
kind/bug severity/P2 status/code-review
**Reproduction Steps:** 1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**. 2. Create any project. 3. Select the project and click "Delete" button . 4. Confirm deleting. **Expected behavior:** The project is removed and not visible in the "Projects" tab. **Observed behavior:** The project is removed but the projects list isn't refreshed. The deleted project name is still visible. **Che version:** 5.3.0-SNAPSHOT **OS and version:** Ubuntu 16.04 LTS **Docker version:** 1.12.3 **Che install:** Docker container **Additional information:** Problem can be reliably reproduced, doesn't happen randomly: [Yes] ![project_refresh](https://cloud.githubusercontent.com/assets/7760565/22877426/06165544-f1de-11e6-81b6-19158c2b1e19.gif)
1.0
The "Projects" info page does not refresh after deleting a project. - **Reproduction Steps:** 1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**. 2. Create any project. 3. Select the project and click "Delete" button . 4. Confirm deleting. **Expected behavior:** The project is removed and not visible in the "Projects" tab. **Observed behavior:** The project is removed but the projects list isn't refreshed. The deleted project name is still visible. **Che version:** 5.3.0-SNAPSHOT **OS and version:** Ubuntu 16.04 LTS **Docker version:** 1.12.3 **Che install:** Docker container **Additional information:** Problem can be reliably reproduced, doesn't happen randomly: [Yes] ![project_refresh](https://cloud.githubusercontent.com/assets/7760565/22877426/06165544-f1de-11e6-81b6-19158c2b1e19.gif)
non_process
the projects info page does not refresh after deleting a project reproduction steps workspaces configure workspace select tab projects create any project select the project and click delete button confirm deleting expected behavior the project is removed and not visible in the projects tab observed behavior the project is removed but the projects list isn t refreshed the deleted project name is still visible che version snapshot os and version ubuntu lts docker version che install docker container additional information problem can be reliably reproduced doesn t happen randomly
0
4,398
7,294,514,949
IssuesEvent
2018-02-26 00:23:15
nyu-software-engineering/online-time-tracker
https://api.github.com/repos/nyu-software-engineering/online-time-tracker
closed
Read Chrome extension documentation and create a trivial extension
2-in process Sprint 0 task
This task is part of User Story #14
1.0
Read Chrome extension documentation and create a trivial extension - This task is part of User Story #14
process
read chrome extension documentation and create a trivial extension this task is part of user story
1
710,754
24,432,293,097
IssuesEvent
2022-10-06 08:59:57
unep-grid/map-x-mgl
https://api.github.com/repos/unep-grid/map-x-mgl
closed
Missing translations & typos [end 2021 update]
priority 1 text/Translation
_This issue is an update of #327_ ### Edit style - [x] change "opacity" to "transparency" to be consistent with the rest of the application ### Area intersection tools - [x] _select analysis_ section is not translated ### Manage sources - [x] pop-up title: Edit source -> Gérer la source / Editar la fuente de datos ### Search engine API - [x] all terms in the panel ### Map composer - [x] all terms in the module ### Draw tool - [x] missing entry for `draw_feature_type` ### Story maps engine - [x] slide -> In Spanish should be "Diapositiva" - [x] slide classes (card, text-right ....) - [x] in edition mode: all tooltips are not translated ### Attribut table - [x] clear filter - [x] export CSV ### Other remarks - [x] ~texts accompanying the progress bars are not translated~ Will be refactored - [x] ~logs are not translated~ - [x] lack of consistency between the terms Configure and Edit e.g. the "Configure view" button becomes "Edit view (in the header).
1.0
Missing translations & typos [end 2021 update] - _This issue is an update of #327_ ### Edit style - [x] change "opacity" to "transparency" to be consistent with the rest of the application ### Area intersection tools - [x] _select analysis_ section is not translated ### Manage sources - [x] pop-up title: Edit source -> Gérer la source / Editar la fuente de datos ### Search engine API - [x] all terms in the panel ### Map composer - [x] all terms in the module ### Draw tool - [x] missing entry for `draw_feature_type` ### Story maps engine - [x] slide -> In Spanish should be "Diapositiva" - [x] slide classes (card, text-right ....) - [x] in edition mode: all tooltips are not translated ### Attribut table - [x] clear filter - [x] export CSV ### Other remarks - [x] ~texts accompanying the progress bars are not translated~ Will be refactored - [x] ~logs are not translated~ - [x] lack of consistency between the terms Configure and Edit e.g. the "Configure view" button becomes "Edit view (in the header).
non_process
missing translations typos this issue is an update of edit style change opacity to transparency to be consistent with the rest of the application area intersection tools select analysis section is not translated manage sources pop up title edit source gérer la source editar la fuente de datos search engine api all terms in the panel map composer all terms in the module draw tool missing entry for draw feature type story maps engine slide in spanish should be diapositiva slide classes card text right in edition mode all tooltips are not translated attribut table clear filter export csv other remarks texts accompanying the progress bars are not translated will be refactored logs are not translated lack of consistency between the terms configure and edit e g the configure view button becomes edit view in the header
0
14,689
17,836,486,711
IssuesEvent
2021-09-03 02:19:57
jim-king-2000/IndustryCamera
https://api.github.com/repos/jim-king-2000/IndustryCamera
closed
[bug]: 多路拉流时,停止一路流,所有流都会停止
bug processing A
### 问题描述 多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断 ### 您预期的行为 停止一路拉流,不影响其他端拉流 ### 系统表现的行为 停止一路拉流,其他端拉流终端 ### 复现路径 开启两个网页端拉流,停止一路,会发现另一路流中断 ### 辅助信息 - 浏览器版本:Edge/Chrome 92 - 固件版本:v1.0
1.0
[bug]: 多路拉流时,停止一路流,所有流都会停止 - ### 问题描述 多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断 ### 您预期的行为 停止一路拉流,不影响其他端拉流 ### 系统表现的行为 停止一路拉流,其他端拉流终端 ### 复现路径 开启两个网页端拉流,停止一路,会发现另一路流中断 ### 辅助信息 - 浏览器版本:Edge/Chrome 92 - 固件版本:v1.0
process
多路拉流时,停止一路流,所有流都会停止 问题描述 多端同时拉取一个设备流时,停止一路拉流,其他路拉流也会中断 您预期的行为 停止一路拉流,不影响其他端拉流 系统表现的行为 停止一路拉流,其他端拉流终端 复现路径 开启两个网页端拉流,停止一路,会发现另一路流中断 辅助信息 浏览器版本:edge chrome 固件版本:
1
15,788
19,977,800,686
IssuesEvent
2022-01-29 11:35:34
bdrum/kaggle
https://api.github.com/repos/bdrum/kaggle
opened
Create pipeline for titanic
enhancement preprocessing model classification pipeline
I have to prepare pipeline for a data that will include such steps as - [ ] feature selection - [ ] preprocessing - [ ] classification
1.0
Create pipeline for titanic - I have to prepare pipeline for a data that will include such steps as - [ ] feature selection - [ ] preprocessing - [ ] classification
process
create pipeline for titanic i have to prepare pipeline for a data that will include such steps as feature selection preprocessing classification
1
11,901
14,697,307,547
IssuesEvent
2021-01-04 02:50:53
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Case sensitive field calculator (using graphical modeler)
Feature Request Modeller Processing
**Feature description.** At the moment, Field Calculator does not seem to be case sensitive when creating a new field. Therefore, it is not possible to simply rename a field for which you can't decide a name (e.g. `Distance` to `distance` created with `Distance matrix`) It would be great to be able to do such action without having to overburden the model (see below). **Additional context** There are some alternatives, as discussed on this SO post : [https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler](https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler) Yet, using the `Refactoring fields` algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler. Currently, the best alternative is to add several steps to the process : create a temporary new field, copy the desired value in this field, then rename the temporary field.
1.0
Case sensitive field calculator (using graphical modeler) - **Feature description.** At the moment, Field Calculator does not seem to be case sensitive when creating a new field. Therefore, it is not possible to simply rename a field for which you can't decide a name (e.g. `Distance` to `distance` created with `Distance matrix`) It would be great to be able to do such action without having to overburden the model (see below). **Additional context** There are some alternatives, as discussed on this SO post : [https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler](https://gis.stackexchange.com/questions/382306/field-calculator-case-sensitive-issue-in-qgis-graphical-modeler) Yet, using the `Refactoring fields` algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler. Currently, the best alternative is to add several steps to the process : create a temporary new field, copy the desired value in this field, then rename the temporary field.
process
case sensitive field calculator using graphical modeler feature description at the moment field calculator does not seem to be case sensitive when creating a new field therefore it is not possible to simply rename a field for which you can t decide a name e g distance to distance created with distance matrix it would be great to be able to do such action without having to overburden the model see below additional context there are some alternatives as discussed on this so post yet using the refactoring fields algorithm requires knowledge of the field names in the input layer so this may not be an option when creating a model in the graphical modeler currently the best alternative is to add several steps to the process create a temporary new field copy the desired value in this field then rename the temporary field
1
3,664
6,694,769,035
IssuesEvent
2017-10-10 04:19:28
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
opened
View Watering Assignment - Display the Watering Assignment
enhancement process workflow
Record the date and time in a new table. Display the date and time associated with assignment number in the select list.
1.0
View Watering Assignment - Display the Watering Assignment - Record the date and time in a new table. Display the date and time associated with assignment number in the select list.
process
view watering assignment display the watering assignment record the date and time in a new table display the date and time associated with assignment number in the select list
1
127,705
27,105,237,104
IssuesEvent
2023-02-15 11:36:21
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Typo in "Your first Flutter app" step 6
d: codelabs documentation
Issue for this codelab: https://codelabs.developers.google.com/codelabs/flutter-codelab-first#5 Very small typo that might, however, take your code out of sync with the example: ``` Next, add the Like button and connect it to toggleFavorites(). For a challenge, first try to do this by yourself, without looking at the code block below. ``` The function name reference here should be `toggleFavorite()`, not Favorites.
1.0
Typo in "Your first Flutter app" step 6 - Issue for this codelab: https://codelabs.developers.google.com/codelabs/flutter-codelab-first#5 Very small typo that might, however, take your code out of sync with the example: ``` Next, add the Like button and connect it to toggleFavorites(). For a challenge, first try to do this by yourself, without looking at the code block below. ``` The function name reference here should be `toggleFavorite()`, not Favorites.
non_process
typo in your first flutter app step issue for this codelab very small typo that might however take your code out of sync with the example next add the like button and connect it to togglefavorites for a challenge first try to do this by yourself without looking at the code block below the function name reference here should be togglefavorite not favorites
0
659,547
21,932,515,066
IssuesEvent
2022-05-23 11:01:41
PerfectFit-project/virtual-coach-issues
https://api.github.com/repos/PerfectFit-project/virtual-coach-issues
closed
Implement part of future-self dialog in botkit
spike high priority
Follow-up from #250. Botkit seems perfect for our type of dialogs. But let's investigate how good the fit is if we want to go into this path.
1.0
Implement part of future-self dialog in botkit - Follow-up from #250. Botkit seems perfect for our type of dialogs. But let's investigate how good the fit is if we want to go into this path.
non_process
implement part of future self dialog in botkit follow up from botkit seems perfect for our type of dialogs but let s investigate how good the fit is if we want to go into this path
0
111,142
17,015,798,746
IssuesEvent
2021-07-02 11:51:49
anyulled/react-playground
https://api.github.com/repos/anyulled/react-playground
opened
CVE-2021-29059 (High) detected in is-svg-3.0.0.tgz
security vulnerability
## CVE-2021-29059 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p> <p>Path to dependency file: react-playground/package.json</p> <p>Path to vulnerable library: react-playground/node_modules/is-svg</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - optimize-css-assets-webpack-plugin-5.0.4.tgz - cssnano-4.1.10.tgz - cssnano-preset-default-4.0.7.tgz - postcss-svgo-4.0.2.tgz - :x: **is-svg-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/anyulled/react-playground/commit/523de5a20017f19a3c0d97781ef28d338d041797">523de5a20017f19a3c0d97781ef28d338d041797</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was discovered in IS-SVG version 4.3.1 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29059 (High) detected in is-svg-3.0.0.tgz - ## CVE-2021-29059 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p> <p>Path to dependency file: react-playground/package.json</p> <p>Path to vulnerable library: react-playground/node_modules/is-svg</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - optimize-css-assets-webpack-plugin-5.0.4.tgz - cssnano-4.1.10.tgz - cssnano-preset-default-4.0.7.tgz - postcss-svgo-4.0.2.tgz - :x: **is-svg-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/anyulled/react-playground/commit/523de5a20017f19a3c0d97781ef28d338d041797">523de5a20017f19a3c0d97781ef28d338d041797</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was discovered in IS-SVG version 4.3.1 and below where a Regular Expression Denial of Service (ReDOS) occurs if the application is provided and checks a crafted invalid SVG string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29059>CVE-2021-29059</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in is svg tgz cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file react playground package json path to vulnerable library react playground node modules is svg dependency hierarchy react scripts tgz root library optimize css assets webpack plugin tgz cssnano tgz cssnano preset default tgz postcss svgo tgz x is svg tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was discovered in is svg version and below where a regular expression denial of service redos occurs if the application is provided and checks a crafted invalid svg string publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0