Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
15,854
11,731,195,188
IssuesEvent
2020-03-10 23:20:17
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Merge multiple input (ie csv) files into a single table in the db
interface/infrastructure newfeature
We need the ability to read multiple .csv files and merge them into a single table in the database.
1.0
Merge multiple input (ie csv) files into a single table in the db - We need the ability to read multiple .csv files and merge them into a single table in the database.
infrastructure
merge multiple input ie csv files into a single table in the db we need the ability to read multiple csv files and merge them into a single table in the database
1
21,989
14,949,818,756
IssuesEvent
2021-01-26 12:08:12
airyhq/airy
https://api.github.com/repos/airyhq/airy
closed
Don't deploy optional Airy components when core starts
infrastructure
The optional Airy components should be deployed with 0 replicas and after that managed by the controller - started if configuration for them is present. Also the configMaps which are particular for those components, shouldn't be deployed upfront, so that the controller can scale the apps up after the configMap is created. At this moment, optional components are all the sources and the webhook integration.
1.0
Don't deploy optional Airy components when core starts - The optional Airy components should be deployed with 0 replicas and after that managed by the controller - started if configuration for them is present. Also the configMaps which are particular for those components, shouldn't be deployed upfront, so that the controller can scale the apps up after the configMap is created. At this moment, optional components are all the sources and the webhook integration.
infrastructure
don t deploy optional airy components when core starts the optional airy components should be deployed with replicas and after that managed by the controller started if configuration for them is present also the configmaps which are particular for those components shouldn t be deployed upfront so that the controller can scale the apps up after the configmap is created at this moment optional components are all the sources and the webhook integration
1
8,424
7,422,762,892
IssuesEvent
2018-03-23 01:12:19
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Tests seem to time out before the default timeout expires
area-Infrastructure test bug
According to: https://github.com/dotnet/coreclr/blob/master/tests/src/Common/Coreclr.TestWrapper/CoreclrTestWrapperLib.cs#L99 The default timeout should be 10 minutes. It doesn't look like the following tests override that default timeout, and yet they seem to be timing out in less than or around one minute in the CI. From under tests/src/baseservices/threading/generics/Monitor: EnterExit14 and waitone2: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs0x1000_prtest/8/ EnterExit14: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs2_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs3_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_tailcallstress_prtest/21/ EnterExit14, EnterExit12: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs4_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstressregs8_prtest/15/
1.0
Tests seem to time out before the default timeout expires - According to: https://github.com/dotnet/coreclr/blob/master/tests/src/Common/Coreclr.TestWrapper/CoreclrTestWrapperLib.cs#L99 The default timeout should be 10 minutes. It doesn't look like the following tests override that default timeout, and yet they seem to be timing out in less than or around one minute in the CI. From under tests/src/baseservices/threading/generics/Monitor: EnterExit14 and waitone2: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs0x1000_prtest/8/ EnterExit14: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs2_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs3_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_tailcallstress_prtest/21/ EnterExit14, EnterExit12: https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstress2_jitstressregs4_prtest/9/ https://ci.dot.net/job/dotnet_coreclr/job/master/job/jitstress/job/x86_arm_altjit_checked_windows_nt_jitstressregs8_prtest/15/
infrastructure
tests seem to time out before the default timeout expires according to the default timeout should be minutes it doesn t look like the following tests override that default timeout and yet they seem to be timing out in less than or around one minute in the ci from under tests src baseservices threading generics monitor and
1
443,787
30,929,032,420
IssuesEvent
2023-08-06 21:17:27
nextauthjs/next-auth
https://api.github.com/repos/nextauthjs/next-auth
closed
Nextjs 13 App directory set cookies
documentation triage
### What is the improvement or update you wish to see? The docs state `Currently, the underlying Next.js cookies() method [only provides read access](https://beta.nextjs.org/docs/api-reference/cookies) to the request cookies. This means that the expires value is stripped away from session in Server Components. Furthermore, there is a hard expiry on sessions, after which the user will be required to sign in again. (The default expiry is 30 days).` However the cookies docs for nextjs that is linked to shows that cookies can be set now. So is this still the case? Is it not possible to avoid the user being signed out when using the app directory? ### Is there any context that might help us understand? Nextjs docs [here](https://nextjs.org/docs/app/api-reference/functions/cookies) ### Does the docs page already exist? Please link to it. https://next-auth.js.org/configuration/nextjs
1.0
Nextjs 13 App directory set cookies - ### What is the improvement or update you wish to see? The docs state `Currently, the underlying Next.js cookies() method [only provides read access](https://beta.nextjs.org/docs/api-reference/cookies) to the request cookies. This means that the expires value is stripped away from session in Server Components. Furthermore, there is a hard expiry on sessions, after which the user will be required to sign in again. (The default expiry is 30 days).` However the cookies docs for nextjs that is linked to shows that cookies can be set now. So is this still the case? Is it not possible to avoid the user being signed out when using the app directory? ### Is there any context that might help us understand? Nextjs docs [here](https://nextjs.org/docs/app/api-reference/functions/cookies) ### Does the docs page already exist? Please link to it. https://next-auth.js.org/configuration/nextjs
non_infrastructure
nextjs app directory set cookies what is the improvement or update you wish to see the docs state currently the underlying next js cookies method to the request cookies this means that the expires value is stripped away from session in server components furthermore there is a hard expiry on sessions after which the user will be required to sign in again the default expiry is days however the cookies docs for nextjs that is linked to shows that cookies can be set now so is this still the case is it not possible to avoid the user being signed out when using the app directory is there any context that might help us understand nextjs docs does the docs page already exist please link to it
0
35,426
31,276,446,440
IssuesEvent
2023-08-22 06:40:59
Yonatan-Lavie/aws-cdk-mongodb-api
https://api.github.com/repos/Yonatan-Lavie/aws-cdk-mongodb-api
closed
create GitHub Actions files
infrastructure
GitHub Actions Create .github/workflows/build.yml. Create .github/workflows/test.yml. Create .github/workflows/deploy-dev.yml. Create .github/workflows/deploy-prod.yml.
1.0
create GitHub Actions files - GitHub Actions Create .github/workflows/build.yml. Create .github/workflows/test.yml. Create .github/workflows/deploy-dev.yml. Create .github/workflows/deploy-prod.yml.
infrastructure
create github actions files github actions create github workflows build yml create github workflows test yml create github workflows deploy dev yml create github workflows deploy prod yml
1
5,116
5,444,868,524
IssuesEvent
2017-03-07 04:54:12
SWE574-Unity/Heritage
https://api.github.com/repos/SWE574-Unity/Heritage
closed
CI/CD pipeline
configuration deployment discussion infrastructure
I want to discuss CI/CD for integration. If we're gonna do it, there are three things to deploy: - [ ] Backend - [ ] Mobile - [ ] Frontend (If we use a seperate view layer, see [#10](https://github.com/SWE574-TeamUnity/RepoUnity/issues/10) ) First thing to discuss is CI service. Travis vs. Jenkins ? Here are my marks. | Jenkins | Travis | | ------ | ------ | | Requires host | Self-host ✔| | We maintain it | Self-maintained ✔| | Familiar ✔| Not as familiar | | Private repo ✔| Public only | | Indirect costs | Free for public ✔| | Integration via Git | Direct integration with GitHub ✔| | deployment script host on Jenkins | deployment script host in Repository ✔| In light of that, I say let's try Travis. It's also very easy to learn and write jobs. It's simply a linux machine. --- For mobile, I think a solution like HockeyApp (free up to 2 mobile apps) would be great. With a simple curl request, we can send our apks directly into a private app market from which we can download our apps. CD for mobile, who doesn't want that? We can also look into HockeyApp's cordova plugin for crash reports, updates etc. but that might be too much to decide now. What do you think @SWE574-TeamUnity/unity ?
1.0
CI/CD pipeline - I want to discuss CI/CD for integration. If we're gonna do it, there are three things to deploy: - [ ] Backend - [ ] Mobile - [ ] Frontend (If we use a seperate view layer, see [#10](https://github.com/SWE574-TeamUnity/RepoUnity/issues/10) ) First thing to discuss is CI service. Travis vs. Jenkins ? Here are my marks. | Jenkins | Travis | | ------ | ------ | | Requires host | Self-host ✔| | We maintain it | Self-maintained ✔| | Familiar ✔| Not as familiar | | Private repo ✔| Public only | | Indirect costs | Free for public ✔| | Integration via Git | Direct integration with GitHub ✔| | deployment script host on Jenkins | deployment script host in Repository ✔| In light of that, I say let's try Travis. It's also very easy to learn and write jobs. It's simply a linux machine. --- For mobile, I think a solution like HockeyApp (free up to 2 mobile apps) would be great. With a simple curl request, we can send our apks directly into a private app market from which we can download our apps. CD for mobile, who doesn't want that? We can also look into HockeyApp's cordova plugin for crash reports, updates etc. but that might be too much to decide now. What do you think @SWE574-TeamUnity/unity ?
infrastructure
ci cd pipeline i want to discuss ci cd for integration if we re gonna do it there are three things to deploy backend mobile frontend if we use a seperate view layer see first thing to discuss is ci service travis vs jenkins here are my marks jenkins travis requires host self host ✔ we maintain it self maintained ✔ familiar ✔ not as familiar private repo ✔ public only indirect costs free for public ✔ integration via git direct integration with github ✔ deployment script host on jenkins deployment script host in repository ✔ in light of that i say let s try travis it s also very easy to learn and write jobs it s simply a linux machine for mobile i think a solution like hockeyapp free up to mobile apps would be great with a simple curl request we can send our apks directly into a private app market from which we can download our apps cd for mobile who doesn t want that we can also look into hockeyapp s cordova plugin for crash reports updates etc but that might be too much to decide now what do you think teamunity unity
1
32,976
27,118,074,583
IssuesEvent
2023-02-15 20:16:20
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Runtime assemblies contain invalid type forwarders
question area-Infrastructure-libraries untriaged
### Description Some runtime assemblies contain type forwarders to other assemblies which are not shipped with the SDK. I noticed this while building a dependency graph of assemblies referenced by my project. `System.IO.Ports` couldn't be found, but it was referenced by `System`. ### Reproduction Steps Run this program, it's a bit sloppy (as it uses `Assembly.LoadFrom` and assumes the file name and assembly name are equal), but it demonstrates the point: ```csharp using System; using System.IO; using System.Linq; using System.Reflection; namespace ConsoleApp; public static class Program { public static void Main() { var dir = Path.GetDirectoryName(typeof(object).Assembly.Location)!; var assemblies = Directory.GetFiles(dir) .Where(f => f.EndsWith(".dll", StringComparison.OrdinalIgnoreCase)) .Select(Path.GetFileNameWithoutExtension) .ToHashSet(); foreach (var assembly in assemblies.Order(StringComparer.OrdinalIgnoreCase)) { try { var asm = Assembly.LoadFrom(Path.Combine(dir, $"{assembly}.dll")); foreach (var refAsmName in asm.GetReferencedAssemblies()) { if (!assemblies.Contains(refAsmName.Name)) Console.WriteLine($"{assembly} -> {refAsmName.Name}"); } } catch { // Native dll } } } } ``` ### Expected behavior This program should produce no output. All references within the SDK should be resolved. ### Actual behavior The following output is produced with .NET 7.0.3: ``` mscorlib -> System.Security.Permissions mscorlib -> System.Threading.AccessControl System -> System.CodeDom System -> Microsoft.Win32.SystemEvents System -> System.Configuration.ConfigurationManager System -> System.Diagnostics.PerformanceCounter System -> System.Diagnostics.EventLog System -> System.Security.Permissions System -> System.IO.Ports System -> System.Windows.Extensions System -> System.Threading.AccessControl System.Configuration -> System.Configuration.ConfigurationManager System.Configuration -> System.Security.Permissions System.Core -> System.Diagnostics.EventLog System.Core -> System.Diagnostics.PerformanceCounter System.Data -> System.Data.SqlClient System.Data -> System.Security.Permissions System.Data -> System.Data.Odbc System.Data -> System.Data.OleDb System.Drawing -> System.Drawing.Common System.Drawing -> System.Configuration.ConfigurationManager System.Drawing -> System.Security.Permissions System.Net -> System.Security.Permissions System.Runtime.Serialization -> System.Runtime.Serialization.Schema System.Security -> System.Security.Cryptography.Pkcs System.Security -> System.Security.Cryptography.ProtectedData System.Security -> System.Windows.Extensions System.Security -> System.Security.Cryptography.Xml System.Security -> System.Security.Permissions System.ServiceModel.Web -> System.ServiceModel.Syndication System.ServiceProcess -> System.ServiceProcess.ServiceController System.ServiceProcess -> System.Security.Permissions System.Transactions -> System.Security.Permissions WindowsBase -> System.IO.Packaging WindowsBase -> System.Security.Permissions ``` The format is `assembly -> referenced assembly`, where `referenced assembly` couldn't be resolved. All of the assemblies listed on the left of the `->` contain only type forwarders. ### Regression? _No response_ ### Known Workarounds _No response_ ### Configuration <details> <summary>dotnet --info</summary> ``` .NET SDK: Version: 7.0.103 Commit: 276c71d299 Runtime Environment: OS Name: Windows OS Version: 10.0.22621 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\7.0.103\ Host: Version: 7.0.3 Architecture: x64 Commit: 0a2bda10e8 .NET SDKs installed: 3.1.426 [C:\Program Files\dotnet\sdk] 5.0.403 [C:\Program Files\dotnet\sdk] 5.0.405 [C:\Program Files\dotnet\sdk] 5.0.408 [C:\Program Files\dotnet\sdk] 6.0.203 [C:\Program Files\dotnet\sdk] 6.0.309 [C:\Program Files\dotnet\sdk] 7.0.100-rc.2.22477.23 [C:\Program Files\dotnet\sdk] 7.0.101 [C:\Program Files\dotnet\sdk] 7.0.102 [C:\Program Files\dotnet\sdk] 7.0.103 [C:\Program Files\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0-rc.2.22476.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0-rc.2.22472.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0-rc.2.22472.13 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Other architectures found: x86 [C:\Program Files (x86)\dotnet] registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation] Environment variables: Not set global.json file: Not found Learn more: https://aka.ms/dotnet/info Download .NET: https://aka.ms/dotnet/download ``` </details> ### Other information _No response_
1.0
Runtime assemblies contain invalid type forwarders - ### Description Some runtime assemblies contain type forwarders to other assemblies which are not shipped with the SDK. I noticed this while building a dependency graph of assemblies referenced by my project. `System.IO.Ports` couldn't be found, but it was referenced by `System`. ### Reproduction Steps Run this program, it's a bit sloppy (as it uses `Assembly.LoadFrom` and assumes the file name and assembly name are equal), but it demonstrates the point: ```csharp using System; using System.IO; using System.Linq; using System.Reflection; namespace ConsoleApp; public static class Program { public static void Main() { var dir = Path.GetDirectoryName(typeof(object).Assembly.Location)!; var assemblies = Directory.GetFiles(dir) .Where(f => f.EndsWith(".dll", StringComparison.OrdinalIgnoreCase)) .Select(Path.GetFileNameWithoutExtension) .ToHashSet(); foreach (var assembly in assemblies.Order(StringComparer.OrdinalIgnoreCase)) { try { var asm = Assembly.LoadFrom(Path.Combine(dir, $"{assembly}.dll")); foreach (var refAsmName in asm.GetReferencedAssemblies()) { if (!assemblies.Contains(refAsmName.Name)) Console.WriteLine($"{assembly} -> {refAsmName.Name}"); } } catch { // Native dll } } } } ``` ### Expected behavior This program should produce no output. All references within the SDK should be resolved. ### Actual behavior The following output is produced with .NET 7.0.3: ``` mscorlib -> System.Security.Permissions mscorlib -> System.Threading.AccessControl System -> System.CodeDom System -> Microsoft.Win32.SystemEvents System -> System.Configuration.ConfigurationManager System -> System.Diagnostics.PerformanceCounter System -> System.Diagnostics.EventLog System -> System.Security.Permissions System -> System.IO.Ports System -> System.Windows.Extensions System -> System.Threading.AccessControl System.Configuration -> System.Configuration.ConfigurationManager System.Configuration -> System.Security.Permissions System.Core -> System.Diagnostics.EventLog System.Core -> System.Diagnostics.PerformanceCounter System.Data -> System.Data.SqlClient System.Data -> System.Security.Permissions System.Data -> System.Data.Odbc System.Data -> System.Data.OleDb System.Drawing -> System.Drawing.Common System.Drawing -> System.Configuration.ConfigurationManager System.Drawing -> System.Security.Permissions System.Net -> System.Security.Permissions System.Runtime.Serialization -> System.Runtime.Serialization.Schema System.Security -> System.Security.Cryptography.Pkcs System.Security -> System.Security.Cryptography.ProtectedData System.Security -> System.Windows.Extensions System.Security -> System.Security.Cryptography.Xml System.Security -> System.Security.Permissions System.ServiceModel.Web -> System.ServiceModel.Syndication System.ServiceProcess -> System.ServiceProcess.ServiceController System.ServiceProcess -> System.Security.Permissions System.Transactions -> System.Security.Permissions WindowsBase -> System.IO.Packaging WindowsBase -> System.Security.Permissions ``` The format is `assembly -> referenced assembly`, where `referenced assembly` couldn't be resolved. All of the assemblies listed on the left of the `->` contain only type forwarders. ### Regression? _No response_ ### Known Workarounds _No response_ ### Configuration <details> <summary>dotnet --info</summary> ``` .NET SDK: Version: 7.0.103 Commit: 276c71d299 Runtime Environment: OS Name: Windows OS Version: 10.0.22621 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\7.0.103\ Host: Version: 7.0.3 Architecture: x64 Commit: 0a2bda10e8 .NET SDKs installed: 3.1.426 [C:\Program Files\dotnet\sdk] 5.0.403 [C:\Program Files\dotnet\sdk] 5.0.405 [C:\Program Files\dotnet\sdk] 5.0.408 [C:\Program Files\dotnet\sdk] 6.0.203 [C:\Program Files\dotnet\sdk] 6.0.309 [C:\Program Files\dotnet\sdk] 7.0.100-rc.2.22477.23 [C:\Program Files\dotnet\sdk] 7.0.101 [C:\Program Files\dotnet\sdk] 7.0.102 [C:\Program Files\dotnet\sdk] 7.0.103 [C:\Program Files\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0-rc.2.22476.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0-rc.2.22472.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.21 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.22 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.32 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.10 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.12 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.14 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 5.0.17 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.5 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.14 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0-rc.2.22472.13 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.3 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Other architectures found: x86 [C:\Program Files (x86)\dotnet] registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation] Environment variables: Not set global.json file: Not found Learn more: https://aka.ms/dotnet/info Download .NET: https://aka.ms/dotnet/download ``` </details> ### Other information _No response_
infrastructure
runtime assemblies contain invalid type forwarders description some runtime assemblies contain type forwarders to other assemblies which are not shipped with the sdk i noticed this while building a dependency graph of assemblies referenced by my project system io ports couldn t be found but it was referenced by system reproduction steps run this program it s a bit sloppy as it uses assembly loadfrom and assumes the file name and assembly name are equal but it demonstrates the point csharp using system using system io using system linq using system reflection namespace consoleapp public static class program public static void main var dir path getdirectoryname typeof object assembly location var assemblies directory getfiles dir where f f endswith dll stringcomparison ordinalignorecase select path getfilenamewithoutextension tohashset foreach var assembly in assemblies order stringcomparer ordinalignorecase try var asm assembly loadfrom path combine dir assembly dll foreach var refasmname in asm getreferencedassemblies if assemblies contains refasmname name console writeline assembly refasmname name catch native dll expected behavior this program should produce no output all references within the sdk should be resolved actual behavior the following output is produced with net mscorlib system security permissions mscorlib system threading accesscontrol system system codedom system microsoft systemevents system system configuration configurationmanager system system diagnostics performancecounter system system diagnostics eventlog system system security permissions system system io ports system system windows extensions system system threading accesscontrol system configuration system configuration configurationmanager system configuration system security permissions system core system diagnostics eventlog system core system diagnostics performancecounter system data system data sqlclient system data system security permissions system data system data odbc system data system data oledb system drawing system drawing common system drawing system configuration configurationmanager system drawing system security permissions system net system security permissions system runtime serialization system runtime serialization schema system security system security cryptography pkcs system security system security cryptography protecteddata system security system windows extensions system security system security cryptography xml system security system security permissions system servicemodel web system servicemodel syndication system serviceprocess system serviceprocess servicecontroller system serviceprocess system security permissions system transactions system security permissions windowsbase system io packaging windowsbase system security permissions the format is assembly referenced assembly where referenced assembly couldn t be resolved all of the assemblies listed on the left of the contain only type forwarders regression no response known workarounds no response configuration dotnet info net sdk version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host version architecture commit net sdks installed rc net runtimes installed microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app rc microsoft aspnetcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app rc microsoft netcore app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app rc microsoft windowsdesktop app other architectures found registered at environment variables not set global json file not found learn more download net other information no response
1
6,007
6,104,394,558
IssuesEvent
2017-06-20 20:52:03
dotnet/wcf
https://api.github.com/repos/dotnet/wcf
opened
Enable all WCF facades for buildtoolsv2
Infrastructure
Facade projects as well as related unit test projects. S.SM.Duplex S.SM.Http S.SM.NetTcp S.SM.Security
1.0
Enable all WCF facades for buildtoolsv2 - Facade projects as well as related unit test projects. S.SM.Duplex S.SM.Http S.SM.NetTcp S.SM.Security
infrastructure
enable all wcf facades for facade projects as well as related unit test projects s sm duplex s sm http s sm nettcp s sm security
1
9,831
8,197,607,954
IssuesEvent
2018-08-31 13:56:13
Tribute-coop/server
https://api.github.com/repos/Tribute-coop/server
opened
Remove Interface and Organization models
Core Deployment Enhancement Infrastructure Python Refactoring
The interface and organization models are no longer required. Domain names will use wildcards to access to packages. To be detailed ....
1.0
Remove Interface and Organization models - The interface and organization models are no longer required. Domain names will use wildcards to access to packages. To be detailed ....
infrastructure
remove interface and organization models the interface and organization models are no longer required domain names will use wildcards to access to packages to be detailed
1
339,024
30,337,451,403
IssuesEvent
2023-07-11 10:28:04
n0-computer/iroh
https://api.github.com/repos/n0-computer/iroh
opened
flakey tests
bug test
We're battling flakey tests for a while now. This is a quick summary of the ones observed in the past 4-5 days. Currently active: - `cli_provide_folder` [seen here](https://github.com/n0-computer/iroh/actions/runs/5516918836/jobs/10058860217) - `test_mesh_network` [seen here](https://github.com/n0-computer/iroh/actions/runs/5513113037/jobs/10050818388) Past few days: - `test_mesh_network` [seen here](https://github.com/n0-computer/iroh/actions/runs/5507072068/jobs/10036549592) [and here](https://github.com/n0-computer/iroh/actions/runs/5484513530/jobs/9992157753) - `multiple_clients` [here](https://github.com/n0-computer/iroh/actions/runs/5486213912/jobs/9996022409) [and here](https://github.com/n0-computer/iroh/actions/runs/5484888292/jobs/9993005557); this was 4 days ago, haven't seen it since
1.0
flakey tests - We're battling flakey tests for a while now. This is a quick summary of the ones observed in the past 4-5 days. Currently active: - `cli_provide_folder` [seen here](https://github.com/n0-computer/iroh/actions/runs/5516918836/jobs/10058860217) - `test_mesh_network` [seen here](https://github.com/n0-computer/iroh/actions/runs/5513113037/jobs/10050818388) Past few days: - `test_mesh_network` [seen here](https://github.com/n0-computer/iroh/actions/runs/5507072068/jobs/10036549592) [and here](https://github.com/n0-computer/iroh/actions/runs/5484513530/jobs/9992157753) - `multiple_clients` [here](https://github.com/n0-computer/iroh/actions/runs/5486213912/jobs/9996022409) [and here](https://github.com/n0-computer/iroh/actions/runs/5484888292/jobs/9993005557); this was 4 days ago, haven't seen it since
non_infrastructure
flakey tests we re battling flakey tests for a while now this is a quick summary of the ones observed in the past days currently active cli provide folder test mesh network past few days test mesh network multiple clients this was days ago haven t seen it since
0
234,712
7,725,304,086
IssuesEvent
2018-05-24 17:32:07
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
opened
Investigate ESRI Client Side Projection
feedback: discussion feedback: proof of concept improvements: enhancement improvements: quality priority: low
Requires https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1153 to happen first. As of version 3.24, the ESRI API now comes with a client side projection engine https://www.esri.com/arcgis-blog/products/js-api-arcgis/mapping/introducing-the-client-side-projection-engine/ Worth taking a look to answer some of the following questions: * Can it handle the equivalent projections that `proj4` does? More? * If we used this, could we drop the need to hit up `epsg.io` for projection codes? * How does the performance compare to `proj4`? * Can it handle a "handmade" formula like `proj4` can? * Does bundling with ESRI make it more difficult to switch to a different API? (lower priority question given the unlikihood) * Will we break libraries like `terraformer-proj4js` that depend on `proj4`?
1.0
Investigate ESRI Client Side Projection - Requires https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1153 to happen first. As of version 3.24, the ESRI API now comes with a client side projection engine https://www.esri.com/arcgis-blog/products/js-api-arcgis/mapping/introducing-the-client-side-projection-engine/ Worth taking a look to answer some of the following questions: * Can it handle the equivalent projections that `proj4` does? More? * If we used this, could we drop the need to hit up `epsg.io` for projection codes? * How does the performance compare to `proj4`? * Can it handle a "handmade" formula like `proj4` can? * Does bundling with ESRI make it more difficult to switch to a different API? (lower priority question given the unlikihood) * Will we break libraries like `terraformer-proj4js` that depend on `proj4`?
non_infrastructure
investigate esri client side projection requires to happen first as of version the esri api now comes with a client side projection engine worth taking a look to answer some of the following questions can it handle the equivalent projections that does more if we used this could we drop the need to hit up epsg io for projection codes how does the performance compare to can it handle a handmade formula like can does bundling with esri make it more difficult to switch to a different api lower priority question given the unlikihood will we break libraries like terraformer that depend on
0
2,133
3,520,644,647
IssuesEvent
2016-01-12 21:44:07
comses/miracle
https://api.github.com/repos/comses/miracle
opened
migrate to sharcnet
infrastructure
migrate deployr, radiant, django apps over to the sharcnet server for less resource contention
1.0
migrate to sharcnet - migrate deployr, radiant, django apps over to the sharcnet server for less resource contention
infrastructure
migrate to sharcnet migrate deployr radiant django apps over to the sharcnet server for less resource contention
1
16,625
12,068,672,357
IssuesEvent
2020-04-16 15:04:42
synfig/synfig
https://api.github.com/repos/synfig/synfig
opened
Use OpenCollective (transparent ledger) for donations/payments
infrastructure
- [x] Register collective - https://opencollective.com/synfig - [x] Redirect incoming funds from Paddle - [ ] Invite core contributors to collective - [ ] Add link to https://github.com/synfig/synfig/blob/master/.github/FUNDING.yml - [ ] Update Donations page - https://www.synfig.org/donate/ - [ ] Add [GitHub Bot](https://github.com/opencollective/opencollective-bot) - [ ] Blog post announcement
1.0
Use OpenCollective (transparent ledger) for donations/payments - - [x] Register collective - https://opencollective.com/synfig - [x] Redirect incoming funds from Paddle - [ ] Invite core contributors to collective - [ ] Add link to https://github.com/synfig/synfig/blob/master/.github/FUNDING.yml - [ ] Update Donations page - https://www.synfig.org/donate/ - [ ] Add [GitHub Bot](https://github.com/opencollective/opencollective-bot) - [ ] Blog post announcement
infrastructure
use opencollective transparent ledger for donations payments register collective redirect incoming funds from paddle invite core contributors to collective add link to update donations page add blog post announcement
1
8,262
7,312,842,216
IssuesEvent
2018-02-28 22:22:50
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
opened
Enable rule CA1063 (Implement IDisposable correctly) for Roslyn.sln
Area-Analyzers Area-Infrastructure
See [documentation](https://docs.microsoft.com/en-us/visualstudio/code-quality/ca1063-implement-idisposable-correctly). This is a non-DFA based rule that has been implemented in the analyzers package for quite some while, but we recently enabled it by default. This has quite a few violations in Roslyn product and test code. Filing this issue to track fixing/suppressing these as a separate PR.
1.0
Enable rule CA1063 (Implement IDisposable correctly) for Roslyn.sln - See [documentation](https://docs.microsoft.com/en-us/visualstudio/code-quality/ca1063-implement-idisposable-correctly). This is a non-DFA based rule that has been implemented in the analyzers package for quite some while, but we recently enabled it by default. This has quite a few violations in Roslyn product and test code. Filing this issue to track fixing/suppressing these as a separate PR.
infrastructure
enable rule implement idisposable correctly for roslyn sln see this is a non dfa based rule that has been implemented in the analyzers package for quite some while but we recently enabled it by default this has quite a few violations in roslyn product and test code filing this issue to track fixing suppressing these as a separate pr
1
18,519
13,046,340,639
IssuesEvent
2020-07-29 08:50:40
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
Multiproxy infrastructure
Priority: P2 infrastructure stake-off-bubble
### Expected Behavior A developer should be able to deploy a testnet with multiple proxies per validator on K8s & VMs ### Current Behavior Only single proxy per validator supported
1.0
Multiproxy infrastructure - ### Expected Behavior A developer should be able to deploy a testnet with multiple proxies per validator on K8s & VMs ### Current Behavior Only single proxy per validator supported
infrastructure
multiproxy infrastructure expected behavior a developer should be able to deploy a testnet with multiple proxies per validator on vms current behavior only single proxy per validator supported
1
19,414
13,230,906,560
IssuesEvent
2020-08-18 10:39:38
Tribler/tribler
https://api.github.com/repos/Tribler/tribler
closed
Separate release URL for Tribler besides GitHub for download statistics
enhancement infrastructure
Right now, we do release via GitHub which is great but we lack some important insights on release statistics. For example, from Github API, - We know how many downloads were done for each of the release artifacts but we don't know the distribution over time. - Daily/Weekly/Monthly downloads data is not available. - Global spread information is also not available. - How long does it take the majority of the users to download/switch to the newly released version? We don't know. I believe release statistics will be useful to better understand our users to some extent and improve our release cycle. Therefore, as the first step towards https://github.com/Tribler/tribler/issues/5510, I propose to develop a separate release website/service/API that shows all the release statistics based on public information. Update: https://releases.tribler.org (available now)
1.0
Separate release URL for Tribler besides GitHub for download statistics - Right now, we do release via GitHub which is great but we lack some important insights on release statistics. For example, from Github API, - We know how many downloads were done for each of the release artifacts but we don't know the distribution over time. - Daily/Weekly/Monthly downloads data is not available. - Global spread information is also not available. - How long does it take the majority of the users to download/switch to the newly released version? We don't know. I believe release statistics will be useful to better understand our users to some extent and improve our release cycle. Therefore, as the first step towards https://github.com/Tribler/tribler/issues/5510, I propose to develop a separate release website/service/API that shows all the release statistics based on public information. Update: https://releases.tribler.org (available now)
infrastructure
separate release url for tribler besides github for download statistics right now we do release via github which is great but we lack some important insights on release statistics for example from github api we know how many downloads were done for each of the release artifacts but we don t know the distribution over time daily weekly monthly downloads data is not available global spread information is also not available how long does it take the majority of the users to download switch to the newly released version we don t know i believe release statistics will be useful to better understand our users to some extent and improve our release cycle therefore as the first step towards i propose to develop a separate release website service api that shows all the release statistics based on public information update available now
1
13,162
10,132,121,004
IssuesEvent
2019-08-01 21:23:58
microsoft/TypeScript
https://api.github.com/repos/microsoft/TypeScript
closed
Tests don't pass when TS is retargeted to `es6` and moduleResolution: `node`
ES6 Infrastructure
In theory, every `node` version we test on supports full es6. However, if `tsconfig-base.json` is edited to retarget TS's build for ES6 environments, many of our services tests fail - this would seem to indicate a deficiency in our analysis somewhere. :(
1.0
Tests don't pass when TS is retargeted to `es6` and moduleResolution: `node` - In theory, every `node` version we test on supports full es6. However, if `tsconfig-base.json` is edited to retarget TS's build for ES6 environments, many of our services tests fail - this would seem to indicate a deficiency in our analysis somewhere. :(
infrastructure
tests don t pass when ts is retargeted to and moduleresolution node in theory every node version we test on supports full however if tsconfig base json is edited to retarget ts s build for environments many of our services tests fail this would seem to indicate a deficiency in our analysis somewhere
1
20,621
14,086,825,030
IssuesEvent
2020-11-05 04:51:12
openemr/openemr
https://api.github.com/repos/openemr/openemr
closed
Speed Up Gulp Build Time
Enhancement Infrastructure
### Describe the problem This is a follow-up issue to the death of superhero and addition of compact themes. When I added compact themes it basically was asking Gulp to compile all the themes one more time, therefore drastically impacting build time. ### Describe the solution you'd like Bootstrap 4 uses modules to separate different items of itself ([here's the code in where they get separated at](https://github.com/twbs/bootstrap/blob/v4.5.2/scss/bootstrap.scss)). What I plan to do for compact themes is when they are activated that they overlay the regular theme and just override it. There are only a few modules that actually get changed with compact themes and some of them aren't needed (e.g. the BS4 reboot defaults aren't needed). If I can cut some of these modules out of the compact themes, build times could improve. Another thing I plan on doing is reorganizing the themes in the `oe-styles` folder to have one file that has all the variables needed to be changed (this is similar to the way I did this in the [colored themes with this file](https://github.com/openemr/openemr/blob/master/interface/themes/colors/utilities/default_variables.scss)). This way any changes down the line are easy and any overrides of these variables can just occur in the individual theme files. I will also be redoing the Manila theme with this issue. The reason for this is that it holds too many code overrides and needs to be simplified a lot. This should help speed up the build time in theory. Lastly, I will be sweeping through all the colored themes and even removing some of them. I might end up consolidating some into a new theme if that is warranted. The goal of this is to reduce the amount of themes as in my opinion we have WAYYY too many themes which only increases the build time.
1.0
Speed Up Gulp Build Time - ### Describe the problem This is a follow-up issue to the death of superhero and addition of compact themes. When I added compact themes it basically was asking Gulp to compile all the themes one more time, therefore drastically impacting build time. ### Describe the solution you'd like Bootstrap 4 uses modules to separate different items of itself ([here's the code in where they get separated at](https://github.com/twbs/bootstrap/blob/v4.5.2/scss/bootstrap.scss)). What I plan to do for compact themes is when they are activated that they overlay the regular theme and just override it. There are only a few modules that actually get changed with compact themes and some of them aren't needed (e.g. the BS4 reboot defaults aren't needed). If I can cut some of these modules out of the compact themes, build times could improve. Another thing I plan on doing is reorganizing the themes in the `oe-styles` folder to have one file that has all the variables needed to be changed (this is similar to the way I did this in the [colored themes with this file](https://github.com/openemr/openemr/blob/master/interface/themes/colors/utilities/default_variables.scss)). This way any changes down the line are easy and any overrides of these variables can just occur in the individual theme files. I will also be redoing the Manila theme with this issue. The reason for this is that it holds too many code overrides and needs to be simplified a lot. This should help speed up the build time in theory. Lastly, I will be sweeping through all the colored themes and even removing some of them. I might end up consolidating some into a new theme if that is warranted. The goal of this is to reduce the amount of themes as in my opinion we have WAYYY too many themes which only increases the build time.
infrastructure
speed up gulp build time describe the problem this is a follow up issue to the death of superhero and addition of compact themes when i added compact themes it basically was asking gulp to compile all the themes one more time therefore drastically impacting build time describe the solution you d like bootstrap uses modules to separate different items of itself what i plan to do for compact themes is when they are activated that they overlay the regular theme and just override it there are only a few modules that actually get changed with compact themes and some of them aren t needed e g the reboot defaults aren t needed if i can cut some of these modules out of the compact themes build times could improve another thing i plan on doing is reorganizing the themes in the oe styles folder to have one file that has all the variables needed to be changed this is similar to the way i did this in the this way any changes down the line are easy and any overrides of these variables can just occur in the individual theme files i will also be redoing the manila theme with this issue the reason for this is that it holds too many code overrides and needs to be simplified a lot this should help speed up the build time in theory lastly i will be sweeping through all the colored themes and even removing some of them i might end up consolidating some into a new theme if that is warranted the goal of this is to reduce the amount of themes as in my opinion we have wayyy too many themes which only increases the build time
1
24,049
16,769,926,639
IssuesEvent
2021-06-14 13:42:43
google-research/kubric
https://api.github.com/repos/google-research/kubric
closed
Configure python linter
Infrastructure
Currently the automatic linting only fails on syntax errors. It should be properly configured to (partially) enforce our style guidelines.
1.0
Configure python linter - Currently the automatic linting only fails on syntax errors. It should be properly configured to (partially) enforce our style guidelines.
infrastructure
configure python linter currently the automatic linting only fails on syntax errors it should be properly configured to partially enforce our style guidelines
1
283,699
24,559,271,602
IssuesEvent
2022-10-12 18:41:15
mozilla-mobile/android-components
https://api.github.com/repos/mozilla-mobile/android-components
closed
Intermittent test failure - mozilla.components.feature.app.links.AppLinksFeatureTest - feature observes app intents when started
🎲 intermittent-test
### Firebase Test Run: https://treeherder.mozilla.org/logviewer?job_id=341442876&repo=android-components ### Stacktrace: ``` SUITE: mozilla.components.feature.app.links.AppLinksFeatureTest TEST: feature observes app intents when started [Robolectric] mozilla.components.feature.app.links.AppLinksFeatureTest.feature observes app intents when started: sdk=28; resources=binary I/MonitoringInstr: Instrumentation started! FAILURE java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertNotNull(Assert.java:712) at org.junit.Assert.assertNotNull(Assert.java:722) at mozilla.components.feature.app.links.AppLinksFeatureTest.feature observes app intents when started(AppLinksFeatureTest.kt:116) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:601) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.internal.SandboxTestRunner$2.evaluate(SandboxTestRunner.java:260) at org.robolectric.internal.SandboxTestRunner.runChild(SandboxTestRunner.java:130) at org.robolectric.internal.SandboxTestRunner.runChild(SandboxTestRunner.java:42) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.robolectric.internal.SandboxTestRunner$1.evaluate(SandboxTestRunner.java:84) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at androidx.test.ext.junit.runners.AndroidJUnit4.run(AndroidJUnit4.java:154) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38) at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62) at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) at org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:119) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164) at org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:414) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) ``` ### Build: https://github.com/mozilla-mobile/android-components/commit/db779f5b6e4223744bceaea166647d2c63ef1844 ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-332)
1.0
Intermittent test failure - mozilla.components.feature.app.links.AppLinksFeatureTest - feature observes app intents when started - ### Firebase Test Run: https://treeherder.mozilla.org/logviewer?job_id=341442876&repo=android-components ### Stacktrace: ``` SUITE: mozilla.components.feature.app.links.AppLinksFeatureTest TEST: feature observes app intents when started [Robolectric] mozilla.components.feature.app.links.AppLinksFeatureTest.feature observes app intents when started: sdk=28; resources=binary I/MonitoringInstr: Instrumentation started! FAILURE java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertNotNull(Assert.java:712) at org.junit.Assert.assertNotNull(Assert.java:722) at mozilla.components.feature.app.links.AppLinksFeatureTest.feature observes app intents when started(AppLinksFeatureTest.kt:116) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:601) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.internal.SandboxTestRunner$2.evaluate(SandboxTestRunner.java:260) at org.robolectric.internal.SandboxTestRunner.runChild(SandboxTestRunner.java:130) at org.robolectric.internal.SandboxTestRunner.runChild(SandboxTestRunner.java:42) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.robolectric.internal.SandboxTestRunner$1.evaluate(SandboxTestRunner.java:84) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at androidx.test.ext.junit.runners.AndroidJUnit4.run(AndroidJUnit4.java:154) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38) at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62) at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) at org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:119) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164) at org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:414) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) ``` ### Build: https://github.com/mozilla-mobile/android-components/commit/db779f5b6e4223744bceaea166647d2c63ef1844 ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-332)
non_infrastructure
intermittent test failure mozilla components feature app links applinksfeaturetest feature observes app intents when started firebase test run stacktrace suite mozilla components feature app links applinksfeaturetest test feature observes app intents when started mozilla components feature app links applinksfeaturetest feature observes app intents when started sdk resources binary i monitoringinstr instrumentation started failure java lang assertionerror at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert assertnotnull assert java at org junit assert assertnotnull assert java at mozilla components feature app links applinksfeaturetest feature observes app intents when started applinksfeaturetest kt at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org robolectric robolectrictestrunner helpertestrunner evaluate robolectrictestrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit rules testwatcher evaluate testwatcher java at org junit rules runrules evaluate runrules java at org robolectric internal sandboxtestrunner evaluate sandboxtestrunner java at org robolectric internal sandboxtestrunner runchild sandboxtestrunner java at org robolectric internal sandboxtestrunner runchild sandboxtestrunner java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org robolectric internal sandboxtestrunner evaluate sandboxtestrunner java at org junit runners parentrunner run parentrunner java at androidx test ext junit runners run java at org gradle api internal tasks testing junit junittestclassexecutor runtestclass junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit abstractjunittestclassprocessor processtestclass abstractjunittestclassprocessor java at org gradle api internal tasks testing suitetestclassprocessor processtestclass suitetestclassprocessor java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch contextclassloaderdispatch dispatch contextclassloaderdispatch java at org gradle internal dispatch proxydispatchadapter dispatchinginvocationhandler invoke proxydispatchadapter java at com sun proxy processtestclass unknown source at org gradle api internal tasks testing worker testworker processtestclass testworker java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal remote internal hub messagehubbackedobjectconnection dispatchwrapper dispatch messagehubbackedobjectconnection java at org gradle internal remote internal hub messagehubbackedobjectconnection dispatchwrapper dispatch messagehubbackedobjectconnection java at org gradle internal remote internal hub messagehub handler run messagehub java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java at java lang thread run thread java build ┆issue is synchronized with this
0
432,185
12,490,264,108
IssuesEvent
2020-05-31 23:05:09
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
email configuration for password reset
:clipboard: Supervisor :crown: Admin :email: Email :raised_hands: Volunteer Priority: High Status: Available help wanted
Part of epics #3, #4, #105 _We are still deciding on which email service to use, but the code for this can be started to the extent possible before that decision is made. We can test email locally._ **What type of user is this for? [volunteer/supervisor/admin/all** all (volunteers most immediately) **Where does/should this occur?** login & user email **Description** When a user clicks [forgot your password](https://casa-r4g-staging.herokuapp.com/users/password/new) and submits "Send me reset password instructions," a reset password email should be triggered and sent by the system to the user email provided. That email should direct to a [future work, not in this issue] password reset page. References: https://github.com/heartcombo/devise/wiki/How-To:-Allow-users-to-edit-their-password **Email Text** Subject line: Create a password & set up your account Hello [volunteer display name], A CASA/Prince George’s County volunteer console account has been created for you. This console is for logging the time you spend and actions you take on your CASA case. You can log activity with your CASA youth, their family members, their foster family or placement, the DSS worker, your Case Supervisor and others associated with your CASA case (such as teachers and therapists). Your console account is associated with this email. If this is not the correct email to use, please stop here and contact your Case Supervisor to change the email address. If you are ready to get started, please set your password. This is the first step to accessing your new volunteer account. [button text = Set your password] Please reach out to your Case Supervisor if you experience any difficulties or have questions. - CASA/Prince George’s County **Screenshots:** Here is what a PG CASA email looks like: ![PG CASA](https://user-images.githubusercontent.com/62810851/80314748-27310180-87a8-11ea-9a8d-08a7d742f1bf.png) And here are some examples of password reset emails. ![ActionID](https://user-images.githubusercontent.com/62810851/80290522-5e48d980-86fa-11ea-9f69-3f8e8d714ee7.jpg) ![opentable](https://user-images.githubusercontent.com/62810851/80290523-60ab3380-86fa-11ea-9ae1-473314488874.jpg) ![spotify](https://user-images.githubusercontent.com/62810851/80290525-6143ca00-86fa-11ea-9953-58a8622275de.jpg) Relates to issues: #22, #23
1.0
email configuration for password reset - Part of epics #3, #4, #105 _We are still deciding on which email service to use, but the code for this can be started to the extent possible before that decision is made. We can test email locally._ **What type of user is this for? [volunteer/supervisor/admin/all** all (volunteers most immediately) **Where does/should this occur?** login & user email **Description** When a user clicks [forgot your password](https://casa-r4g-staging.herokuapp.com/users/password/new) and submits "Send me reset password instructions," a reset password email should be triggered and sent by the system to the user email provided. That email should direct to a [future work, not in this issue] password reset page. References: https://github.com/heartcombo/devise/wiki/How-To:-Allow-users-to-edit-their-password **Email Text** Subject line: Create a password & set up your account Hello [volunteer display name], A CASA/Prince George’s County volunteer console account has been created for you. This console is for logging the time you spend and actions you take on your CASA case. You can log activity with your CASA youth, their family members, their foster family or placement, the DSS worker, your Case Supervisor and others associated with your CASA case (such as teachers and therapists). Your console account is associated with this email. If this is not the correct email to use, please stop here and contact your Case Supervisor to change the email address. If you are ready to get started, please set your password. This is the first step to accessing your new volunteer account. [button text = Set your password] Please reach out to your Case Supervisor if you experience any difficulties or have questions. - CASA/Prince George’s County **Screenshots:** Here is what a PG CASA email looks like: ![PG CASA](https://user-images.githubusercontent.com/62810851/80314748-27310180-87a8-11ea-9a8d-08a7d742f1bf.png) And here are some examples of password reset emails. ![ActionID](https://user-images.githubusercontent.com/62810851/80290522-5e48d980-86fa-11ea-9f69-3f8e8d714ee7.jpg) ![opentable](https://user-images.githubusercontent.com/62810851/80290523-60ab3380-86fa-11ea-9ae1-473314488874.jpg) ![spotify](https://user-images.githubusercontent.com/62810851/80290525-6143ca00-86fa-11ea-9953-58a8622275de.jpg) Relates to issues: #22, #23
non_infrastructure
email configuration for password reset part of epics we are still deciding on which email service to use but the code for this can be started to the extent possible before that decision is made we can test email locally what type of user is this for volunteer supervisor admin all all volunteers most immediately where does should this occur login user email description when a user clicks and submits send me reset password instructions a reset password email should be triggered and sent by the system to the user email provided that email should direct to a password reset page references email text subject line create a password set up your account hello a casa prince george’s county volunteer console account has been created for you this console is for logging the time you spend and actions you take on your casa case you can log activity with your casa youth their family members their foster family or placement the dss worker your case supervisor and others associated with your casa case such as teachers and therapists your console account is associated with this email if this is not the correct email to use please stop here and contact your case supervisor to change the email address if you are ready to get started please set your password this is the first step to accessing your new volunteer account please reach out to your case supervisor if you experience any difficulties or have questions casa prince george’s county screenshots here is what a pg casa email looks like and here are some examples of password reset emails relates to issues
0
33,831
27,852,048,441
IssuesEvent
2023-03-20 19:29:52
dart-lang/site-www
https://api.github.com/repos/dart-lang/site-www
opened
Consider adding a check for git merge remnants
infrastructure p3-low e1-hours
### Describe the problem Recently in https://github.com/dart-lang/site-www/commits/main, remnants of a broken git merge/rebase were accidentally included with the other changes, along with some broken text. While usually this usually will be avoided or caught, it's not great to let those breakages in by accident, causing confusing or broken text. ### Expected fix Add a CI check for common remnants of a broken or failed merged. ### Additional context _No response_
1.0
Consider adding a check for git merge remnants - ### Describe the problem Recently in https://github.com/dart-lang/site-www/commits/main, remnants of a broken git merge/rebase were accidentally included with the other changes, along with some broken text. While usually this usually will be avoided or caught, it's not great to let those breakages in by accident, causing confusing or broken text. ### Expected fix Add a CI check for common remnants of a broken or failed merged. ### Additional context _No response_
infrastructure
consider adding a check for git merge remnants describe the problem recently in remnants of a broken git merge rebase were accidentally included with the other changes along with some broken text while usually this usually will be avoided or caught it s not great to let those breakages in by accident causing confusing or broken text expected fix add a ci check for common remnants of a broken or failed merged additional context no response
1
161,785
6,136,409,094
IssuesEvent
2017-06-26 09:18:22
openshift/origin
https://api.github.com/repos/openshift/origin
closed
`openshift ex registry` and similar should enable easy updating
component/cli component/imageregistry kind/enhancement priority/P3
Should someone make a mistake, like say not adding the right --images when creating the infrastructure component, it should be easy and obvious how to back that out. Right now, for the registry, it means deleting the DC, the service, and the pod, and then _restarting the master_ to clear the service cache if you tried to use it at any point before realizing your mistake. This is far from obvious to the uninitiated. It should at least be possible to display what was created so the user can figure it out for themself. Better if there is a --delete or --recreate option (which also handles the service cache). Of course, if it were possible to recreate without deleting the service, it wouldn't be necessary to clear that cache.
1.0
`openshift ex registry` and similar should enable easy updating - Should someone make a mistake, like say not adding the right --images when creating the infrastructure component, it should be easy and obvious how to back that out. Right now, for the registry, it means deleting the DC, the service, and the pod, and then _restarting the master_ to clear the service cache if you tried to use it at any point before realizing your mistake. This is far from obvious to the uninitiated. It should at least be possible to display what was created so the user can figure it out for themself. Better if there is a --delete or --recreate option (which also handles the service cache). Of course, if it were possible to recreate without deleting the service, it wouldn't be necessary to clear that cache.
non_infrastructure
openshift ex registry and similar should enable easy updating should someone make a mistake like say not adding the right images when creating the infrastructure component it should be easy and obvious how to back that out right now for the registry it means deleting the dc the service and the pod and then restarting the master to clear the service cache if you tried to use it at any point before realizing your mistake this is far from obvious to the uninitiated it should at least be possible to display what was created so the user can figure it out for themself better if there is a delete or recreate option which also handles the service cache of course if it were possible to recreate without deleting the service it wouldn t be necessary to clear that cache
0
14,394
10,798,499,361
IssuesEvent
2019-11-06 10:09:56
aarhusstadsarkiv/digiarch
https://api.github.com/repos/aarhusstadsarkiv/digiarch
closed
Consider a package rename
infrastructure
<!-- Hi! :) If applicable, please link guides/articles when submitting infrastructure issues. The markdown syntax for adding links to text is `[text](url)` --> It would be nice if the package was just called `digiarch`:)
1.0
Consider a package rename - <!-- Hi! :) If applicable, please link guides/articles when submitting infrastructure issues. The markdown syntax for adding links to text is `[text](url)` --> It would be nice if the package was just called `digiarch`:)
infrastructure
consider a package rename hi if applicable please link guides articles when submitting infrastructure issues the markdown syntax for adding links to text is url it would be nice if the package was just called digiarch
1
4,786
5,276,982,350
IssuesEvent
2017-02-07 01:14:15
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Resurrect CoreFX Validation on Nano
area-Infrastructure blocking-release
Needs to be enabled, at minimum, in CI. For completeness (as we do for supported platforms), it needs to be enabled in Helix as well.
1.0
Resurrect CoreFX Validation on Nano - Needs to be enabled, at minimum, in CI. For completeness (as we do for supported platforms), it needs to be enabled in Helix as well.
infrastructure
resurrect corefx validation on nano needs to be enabled at minimum in ci for completeness as we do for supported platforms it needs to be enabled in helix as well
1
33,016
27,144,656,036
IssuesEvent
2023-02-16 18:56:42
consensusnetworks/casimir
https://api.github.com/repos/consensusnetworks/casimir
opened
Submodules sometimes are not pulled
infrastructure
Right now we run the pulling command `git submodule update --init --recursive` using a Husky post-merge hook. Will probably fix by adding one more hook that runs on a newly cloned repository when the post-merge hook does not run. This will replace the need to have an additional pull in the local dev script.
1.0
Submodules sometimes are not pulled - Right now we run the pulling command `git submodule update --init --recursive` using a Husky post-merge hook. Will probably fix by adding one more hook that runs on a newly cloned repository when the post-merge hook does not run. This will replace the need to have an additional pull in the local dev script.
infrastructure
submodules sometimes are not pulled right now we run the pulling command git submodule update init recursive using a husky post merge hook will probably fix by adding one more hook that runs on a newly cloned repository when the post merge hook does not run this will replace the need to have an additional pull in the local dev script
1
6,734
9,856,935,128
IssuesEvent
2019-06-20 00:16:40
natario1/CameraView
https://api.github.com/repos/natario1/CameraView
closed
Size.getWidth() NullPointerException
about:frame processing is:question needs:info status:stale
Device: AVD Nexus 6 API 26 CameraView version: 2.0.0-beta04 I've been getting the error below with a high frequency. It's something that I cannot simulate or force to happen, it simply 'happens' without any change in the code. Error log: ``` 2019-04-27 10:18:26.011 24639-24659/com.smartnsens.opencvapp D/Camera: app passed NULL surface 2019-04-27 10:18:26.041 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.075 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.174 24639-24660/com.smartnsens.opencvapp D/skia: onFlyCompress 2019-04-27 10:18:26.231 24639-24660/com.smartnsens.opencvapp E/AndroidRuntime: FATAL EXCEPTION: FrameProcessorsWorker Process: com.smartnsens.opencvapp, PID: 24639 java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.Size.getWidth()' on a null object reference at com.smartnsens.opencvapp.MainActivity$1.process(MainActivity.java:134) at com.otaliastudios.cameraview.CameraView$Callbacks$11.run(CameraView.java:1809) at android.os.Handler.handleCallback(Handler.java:789) at android.os.Handler.dispatchMessage(Handler.java:98) at android.os.Looper.loop(Looper.java:164) at android.os.HandlerThread.run(HandlerThread.java:65) 2019-04-27 10:18:26.255 24639-24659/com.smartnsens.opencvapp D/Camera: app passed NULL surface 2019-04-27 10:18:26.501 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.793 24639-24639/com.smartnsens.opencvapp E/libEGL: call to OpenGL ES API with no current context (logged once per thread) ``` And my FrameProcessor code: ```java camera.addFrameProcessor(new FrameProcessor() { @Override @WorkerThread public void process(Frame frame) { // Get all the frame's data byte[] data = frame.getData(); int rotation = frame.getRotation(); long time = frame.getTime(); Size size = frame.getSize(); int format = frame.getFormat(); // Preview started (first) if (mBitmap == null) { mFrameSize = size.getWidth() * size.getHeight(); mRGBA = new int[mFrameSize]; mBitmap = Bitmap.createBitmap(size.getWidth(), size.getHeight(), Bitmap.Config.ARGB_8888); } // Process the new frame only if the previous DoPreviewFrame() has already finished // the last frame processing and its rendering. if ( !bProcessing ) { int[] rgba = mRGBA; // FPS counter (uses a 20frames trailing window to compute) thisTime = System.currentTimeMillis(); mLastFramesTimes.add(thisTime); // Hold the first 30 frames timestamps before start to process the frames if (mLastFramesTimes.size() <= 30) return; // Encode the raw data frame in a Bitmap image ByteArrayOutputStream out = new ByteArrayOutputStream(); YuvImage yuvImage = new YuvImage(data, ImageFormat.NV21, size.getWidth(), size.getHeight(), null); yuvImage.compressToJpeg(new Rect(0, 0, size.getWidth(), size.getHeight()), 90, out); byte[] imageBytes = out.toByteArray(); bmpImage = CameraUtils.decodeBitmap(imageBytes); mHandler.post(DoPreviewFrame); } } }); } ``` Best Regards. Kleyson Rios.
1.0
Size.getWidth() NullPointerException - Device: AVD Nexus 6 API 26 CameraView version: 2.0.0-beta04 I've been getting the error below with a high frequency. It's something that I cannot simulate or force to happen, it simply 'happens' without any change in the code. Error log: ``` 2019-04-27 10:18:26.011 24639-24659/com.smartnsens.opencvapp D/Camera: app passed NULL surface 2019-04-27 10:18:26.041 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.075 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.174 24639-24660/com.smartnsens.opencvapp D/skia: onFlyCompress 2019-04-27 10:18:26.231 24639-24660/com.smartnsens.opencvapp E/AndroidRuntime: FATAL EXCEPTION: FrameProcessorsWorker Process: com.smartnsens.opencvapp, PID: 24639 java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.Size.getWidth()' on a null object reference at com.smartnsens.opencvapp.MainActivity$1.process(MainActivity.java:134) at com.otaliastudios.cameraview.CameraView$Callbacks$11.run(CameraView.java:1809) at android.os.Handler.handleCallback(Handler.java:789) at android.os.Handler.dispatchMessage(Handler.java:98) at android.os.Looper.loop(Looper.java:164) at android.os.HandlerThread.run(HandlerThread.java:65) 2019-04-27 10:18:26.255 24639-24659/com.smartnsens.opencvapp D/Camera: app passed NULL surface 2019-04-27 10:18:26.501 24639-24663/com.smartnsens.opencvapp D/EGL_emulation: eglMakeCurrent: 0xa86052a0: ver 2 0 (tinfo 0xa8603380) 2019-04-27 10:18:26.793 24639-24639/com.smartnsens.opencvapp E/libEGL: call to OpenGL ES API with no current context (logged once per thread) ``` And my FrameProcessor code: ```java camera.addFrameProcessor(new FrameProcessor() { @Override @WorkerThread public void process(Frame frame) { // Get all the frame's data byte[] data = frame.getData(); int rotation = frame.getRotation(); long time = frame.getTime(); Size size = frame.getSize(); int format = frame.getFormat(); // Preview started (first) if (mBitmap == null) { mFrameSize = size.getWidth() * size.getHeight(); mRGBA = new int[mFrameSize]; mBitmap = Bitmap.createBitmap(size.getWidth(), size.getHeight(), Bitmap.Config.ARGB_8888); } // Process the new frame only if the previous DoPreviewFrame() has already finished // the last frame processing and its rendering. if ( !bProcessing ) { int[] rgba = mRGBA; // FPS counter (uses a 20frames trailing window to compute) thisTime = System.currentTimeMillis(); mLastFramesTimes.add(thisTime); // Hold the first 30 frames timestamps before start to process the frames if (mLastFramesTimes.size() <= 30) return; // Encode the raw data frame in a Bitmap image ByteArrayOutputStream out = new ByteArrayOutputStream(); YuvImage yuvImage = new YuvImage(data, ImageFormat.NV21, size.getWidth(), size.getHeight(), null); yuvImage.compressToJpeg(new Rect(0, 0, size.getWidth(), size.getHeight()), 90, out); byte[] imageBytes = out.toByteArray(); bmpImage = CameraUtils.decodeBitmap(imageBytes); mHandler.post(DoPreviewFrame); } } }); } ``` Best Regards. Kleyson Rios.
non_infrastructure
size getwidth nullpointerexception device avd nexus api cameraview version i ve been getting the error below with a high frequency it s something that i cannot simulate or force to happen it simply happens without any change in the code error log com smartnsens opencvapp d camera app passed null surface com smartnsens opencvapp d egl emulation eglmakecurrent ver tinfo com smartnsens opencvapp d egl emulation eglmakecurrent ver tinfo com smartnsens opencvapp d skia onflycompress com smartnsens opencvapp e androidruntime fatal exception frameprocessorsworker process com smartnsens opencvapp pid java lang nullpointerexception attempt to invoke virtual method int com otaliastudios cameraview size getwidth on a null object reference at com smartnsens opencvapp mainactivity process mainactivity java at com otaliastudios cameraview cameraview callbacks run cameraview java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android os handlerthread run handlerthread java com smartnsens opencvapp d camera app passed null surface com smartnsens opencvapp d egl emulation eglmakecurrent ver tinfo com smartnsens opencvapp e libegl call to opengl es api with no current context logged once per thread and my frameprocessor code java camera addframeprocessor new frameprocessor override workerthread public void process frame frame get all the frame s data byte data frame getdata int rotation frame getrotation long time frame gettime size size frame getsize int format frame getformat preview started first if mbitmap null mframesize size getwidth size getheight mrgba new int mbitmap bitmap createbitmap size getwidth size getheight bitmap config argb process the new frame only if the previous dopreviewframe has already finished the last frame processing and its rendering if bprocessing int rgba mrgba fps counter uses a trailing window to compute thistime system currenttimemillis mlastframestimes add thistime hold the first frames timestamps before start to process the frames if mlastframestimes size return encode the raw data frame in a bitmap image bytearrayoutputstream out new bytearrayoutputstream yuvimage yuvimage new yuvimage data imageformat size getwidth size getheight null yuvimage compresstojpeg new rect size getwidth size getheight out byte imagebytes out tobytearray bmpimage camerautils decodebitmap imagebytes mhandler post dopreviewframe best regards kleyson rios
0
787,711
27,728,461,000
IssuesEvent
2023-03-15 05:32:47
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
opened
Add site banner for March 20th Legal Oral Hearings
Type: Feature Request Priority: 1 1-off tasks Lead: @mekarpeles Affects: Communications
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Internal staff planning details available at https://docs.google.com/document/d/12X3MGLNV2k8w4eMXRd4MbP5QLb70PHuIKNeZ7B_YAQ8/edit#heading=h.wa34425eph6m To be **patch deployed** day before/on March 20th Relates to #7345 (if possible)
1.0
Add site banner for March 20th Legal Oral Hearings - <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Internal staff planning details available at https://docs.google.com/document/d/12X3MGLNV2k8w4eMXRd4MbP5QLb70PHuIKNeZ7B_YAQ8/edit#heading=h.wa34425eph6m To be **patch deployed** day before/on March 20th Relates to #7345 (if possible)
non_infrastructure
add site banner for march legal oral hearings internal staff planning details available at to be patch deployed day before on march relates to if possible
0
19,651
13,349,482,890
IssuesEvent
2020-08-30 01:09:51
KaTeX/KaTeX
https://api.github.com/repos/KaTeX/KaTeX
closed
Lint all js files
good first issue infrastructure
Currently, files to run `eslint` on are specified using globs in the `package.json` script and `.eslintignore`: https://github.com/KaTeX/KaTeX/blob/981e3ee0eb737609e99635b77c94271b2367c9b0/package.json#L87 https://github.com/KaTeX/KaTeX/blob/981e3ee0eb737609e99635b77c94271b2367c9b0/.eslintignore#L1-L6 This prevents running `eslint` out of the box, e.g., `eslint .` (in the root) or in IDE. The only file that is not included in the script but not ignored by `.eslintignore` is `flow-typed` directory. So if we add `flow-typed/npm/*` to `.eslintignore`: ```diff + flow-typed/npm/* ``` we can change the script to: ```diff - "test:lint:js": "eslint *.js src static test contrib dockers website", + "test:lint:js": "eslint .",
1.0
Lint all js files - Currently, files to run `eslint` on are specified using globs in the `package.json` script and `.eslintignore`: https://github.com/KaTeX/KaTeX/blob/981e3ee0eb737609e99635b77c94271b2367c9b0/package.json#L87 https://github.com/KaTeX/KaTeX/blob/981e3ee0eb737609e99635b77c94271b2367c9b0/.eslintignore#L1-L6 This prevents running `eslint` out of the box, e.g., `eslint .` (in the root) or in IDE. The only file that is not included in the script but not ignored by `.eslintignore` is `flow-typed` directory. So if we add `flow-typed/npm/*` to `.eslintignore`: ```diff + flow-typed/npm/* ``` we can change the script to: ```diff - "test:lint:js": "eslint *.js src static test contrib dockers website", + "test:lint:js": "eslint .",
infrastructure
lint all js files currently files to run eslint on are specified using globs in the package json script and eslintignore this prevents running eslint out of the box e g eslint in the root or in ide the only file that is not included in the script but not ignored by eslintignore is flow typed directory so if we add flow typed npm to eslintignore diff flow typed npm we can change the script to diff test lint js eslint js src static test contrib dockers website test lint js eslint
1
12,308
9,683,417,506
IssuesEvent
2019-05-23 11:29:26
cp-ss2019/decentralized-energy-trading
https://api.github.com/repos/cp-ss2019/decentralized-energy-trading
closed
Style of code documentation
infrastructure question
As mentioned in pull-request #22. It would be great if we have a common style of (in-line and method/function) code documentation in the project. As @dakingha69 mentioned, one suggestion would be JSDoc.
1.0
Style of code documentation - As mentioned in pull-request #22. It would be great if we have a common style of (in-line and method/function) code documentation in the project. As @dakingha69 mentioned, one suggestion would be JSDoc.
infrastructure
style of code documentation as mentioned in pull request it would be great if we have a common style of in line and method function code documentation in the project as mentioned one suggestion would be jsdoc
1
101,805
21,788,940,141
IssuesEvent
2022-05-14 15:44:54
dmwm/CRABServer
https://api.github.com/repos/dmwm/CRABServer
opened
remove task blacklisting from Publisher ?
Priority: High Area: Code Cleanup
latest Publisher version in v3.,220514 only uses new API and it is runing well. But still has codes to blacklist tasks in case FMD retrieve takes too long or timeout, that's a bit excessive, timeout may occur for unrelated reasons. Once sure that things are stable, we should remove that code. Or keep it there and introduce an option in config. file to enable it, in case we have problems.
1.0
remove task blacklisting from Publisher ? - latest Publisher version in v3.,220514 only uses new API and it is runing well. But still has codes to blacklist tasks in case FMD retrieve takes too long or timeout, that's a bit excessive, timeout may occur for unrelated reasons. Once sure that things are stable, we should remove that code. Or keep it there and introduce an option in config. file to enable it, in case we have problems.
non_infrastructure
remove task blacklisting from publisher latest publisher version in only uses new api and it is runing well but still has codes to blacklist tasks in case fmd retrieve takes too long or timeout that s a bit excessive timeout may occur for unrelated reasons once sure that things are stable we should remove that code or keep it there and introduce an option in config file to enable it in case we have problems
0
3,708
4,495,068,676
IssuesEvent
2016-08-31 08:52:55
dinyar/uGMTfirmware
https://api.github.com/repos/dinyar/uGMTfirmware
opened
Make it possible to use relative paths in makeProject.sh
enhancement infrastructure
It would be convenient to use relative paths to indicate where the mp7fw should be checked out into.
1.0
Make it possible to use relative paths in makeProject.sh - It would be convenient to use relative paths to indicate where the mp7fw should be checked out into.
infrastructure
make it possible to use relative paths in makeproject sh it would be convenient to use relative paths to indicate where the should be checked out into
1
30,700
25,005,396,078
IssuesEvent
2022-11-03 11:24:42
beamer-bridge/beamer
https://api.github.com/repos/beamer-bridge/beamer
closed
Build a Subgraph for Beamer Contracts
infrastructure :hammer: research :nerd_face:
Build a Subgraph and upload it to the hosting service of the graph We have an already existing implementation in the company which is very close to what we need in the beginning. ### Scope The subgraph will become a powerful tool which lots of projects in the Beamer ecosystem will be able to use. For now which will keep the scope as tiny as we need. We will need the subgraph for #890 which is primarily there to monitor and alert upon any critical behavior regarding claims - false claims - during challenges the false outcome is winning ### Events to capture This leads to two subgraph implementations which will only handle ClaimMade events and RequestFilled events respectively ### Setting up a Subgraph - subgraph.yaml - handlers - schemas We can almost clone our internal repo and replace it with the things we need.
1.0
Build a Subgraph for Beamer Contracts - Build a Subgraph and upload it to the hosting service of the graph We have an already existing implementation in the company which is very close to what we need in the beginning. ### Scope The subgraph will become a powerful tool which lots of projects in the Beamer ecosystem will be able to use. For now which will keep the scope as tiny as we need. We will need the subgraph for #890 which is primarily there to monitor and alert upon any critical behavior regarding claims - false claims - during challenges the false outcome is winning ### Events to capture This leads to two subgraph implementations which will only handle ClaimMade events and RequestFilled events respectively ### Setting up a Subgraph - subgraph.yaml - handlers - schemas We can almost clone our internal repo and replace it with the things we need.
infrastructure
build a subgraph for beamer contracts build a subgraph and upload it to the hosting service of the graph we have an already existing implementation in the company which is very close to what we need in the beginning scope the subgraph will become a powerful tool which lots of projects in the beamer ecosystem will be able to use for now which will keep the scope as tiny as we need we will need the subgraph for which is primarily there to monitor and alert upon any critical behavior regarding claims false claims during challenges the false outcome is winning events to capture this leads to two subgraph implementations which will only handle claimmade events and requestfilled events respectively setting up a subgraph subgraph yaml handlers schemas we can almost clone our internal repo and replace it with the things we need
1
9,078
7,805,503,057
IssuesEvent
2018-06-11 11:04:47
SciTools/iris
https://api.github.com/repos/SciTools/iris
closed
DeprecationWarning when importing iris
SemVer: Minor Type: Infrastructure
I encountered this issue when attempting to get xarray's test suite working with `-W error`: ``` In [1]: import warnings In [2]: warnings.filterwarnings('always') In [3]: import iris /Users/shoyer/miniconda3/envs/xarray-py36/lib/python3.6/site-packages/iris/config.py:94: DeprecationWarning: The SafeConfigParser class has been renamed to ConfigParser in Python 3.2. This alias will be removed in future versions. Use ConfigParser directly instead. config = configparser.SafeConfigParser() /Users/shoyer/miniconda3/envs/xarray-py36/lib/python3.6/importlib/_bootstrap.py:219: ImportWarning: can't resolve package from __spec__ or __package__, falling back on __name__ and __path__ return f(*args, **kwds) [many duplicate lines omitted] ``` I have Iris 2.0.0 installed from conda-forge on Python 3.6.
1.0
DeprecationWarning when importing iris - I encountered this issue when attempting to get xarray's test suite working with `-W error`: ``` In [1]: import warnings In [2]: warnings.filterwarnings('always') In [3]: import iris /Users/shoyer/miniconda3/envs/xarray-py36/lib/python3.6/site-packages/iris/config.py:94: DeprecationWarning: The SafeConfigParser class has been renamed to ConfigParser in Python 3.2. This alias will be removed in future versions. Use ConfigParser directly instead. config = configparser.SafeConfigParser() /Users/shoyer/miniconda3/envs/xarray-py36/lib/python3.6/importlib/_bootstrap.py:219: ImportWarning: can't resolve package from __spec__ or __package__, falling back on __name__ and __path__ return f(*args, **kwds) [many duplicate lines omitted] ``` I have Iris 2.0.0 installed from conda-forge on Python 3.6.
infrastructure
deprecationwarning when importing iris i encountered this issue when attempting to get xarray s test suite working with w error in import warnings in warnings filterwarnings always in import iris users shoyer envs xarray lib site packages iris config py deprecationwarning the safeconfigparser class has been renamed to configparser in python this alias will be removed in future versions use configparser directly instead config configparser safeconfigparser users shoyer envs xarray lib importlib bootstrap py importwarning can t resolve package from spec or package falling back on name and path return f args kwds i have iris installed from conda forge on python
1
9,564
8,034,360,141
IssuesEvent
2018-07-29 17:44:08
Microsoft/visualfsharp
https://api.github.com/repos/Microsoft/visualfsharp
closed
Move to AsyncPackage
Area-Infrastructure Feature Request
Occasionally while using `VisualStudioOpenSource.vsix` I receive notifications indicating that F# Tools are slowing down solution load: ![image](https://cloud.githubusercontent.com/assets/1249087/23808531/0abb8c2c-057f-11e7-8c86-b7a9eb0101ab.png) In the performance manager: ![image](https://cloud.githubusercontent.com/assets/1249087/23808542/1a80d91e-057f-11e7-9e7e-0257d734625d.png) This notification is from a VM so I don't expect this to be typical performance, but it's still something we could improve. I believe this can be mitigated by using an `AsyncPackage` instead of `Package` to load F# tools. For more see: - [How to: Use AsyncPackage to Load VSPackages in the Background](https://msdn.microsoft.com/en-us/library/mt750288.aspx) - [How to: Provide an Asynchronous Visual Studio Service](https://msdn.microsoft.com/en-us/library/mt709265.aspx)
1.0
Move to AsyncPackage - Occasionally while using `VisualStudioOpenSource.vsix` I receive notifications indicating that F# Tools are slowing down solution load: ![image](https://cloud.githubusercontent.com/assets/1249087/23808531/0abb8c2c-057f-11e7-8c86-b7a9eb0101ab.png) In the performance manager: ![image](https://cloud.githubusercontent.com/assets/1249087/23808542/1a80d91e-057f-11e7-9e7e-0257d734625d.png) This notification is from a VM so I don't expect this to be typical performance, but it's still something we could improve. I believe this can be mitigated by using an `AsyncPackage` instead of `Package` to load F# tools. For more see: - [How to: Use AsyncPackage to Load VSPackages in the Background](https://msdn.microsoft.com/en-us/library/mt750288.aspx) - [How to: Provide an Asynchronous Visual Studio Service](https://msdn.microsoft.com/en-us/library/mt709265.aspx)
infrastructure
move to asyncpackage occasionally while using visualstudioopensource vsix i receive notifications indicating that f tools are slowing down solution load in the performance manager this notification is from a vm so i don t expect this to be typical performance but it s still something we could improve i believe this can be mitigated by using an asyncpackage instead of package to load f tools for more see
1
2,945
3,973,594,627
IssuesEvent
2016-05-04 19:08:23
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Win7 builds failing with ""GetTargetMachineInfo" task failed unexpectedly"
blocking-clean-ci Infrastructure
e.g. http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/141/console ``` d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: The "GetTargetMachineInfo" task failed unexpectedly. [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: System.DllNotFoundException: Unable to load DLL 'api-ms-win-core-sysinfo-l1-2-0.dll': The specified module could not be found. (Exception from HRESULT: 0x8007007E) [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Interop.mincore.GetNativeSystemInfo(SYSTEM_INFO& lpSystemInfo) [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at System.Runtime.InteropServices.RuntimeInformation.get_OSArchitecture() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.DotNet.Build.Tasks.GetTargetMachineInfo.Execute() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.Build.BackEnd.TaskExecutionHost.Microsoft.Build.BackEnd.ITaskExecutionHost.Execute() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.Build.BackEnd.TaskBuilder.<ExecuteInstantiatedTask>d__26.MoveNext() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] ```
1.0
Win7 builds failing with ""GetTargetMachineInfo" task failed unexpectedly" - e.g. http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/141/console ``` d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: The "GetTargetMachineInfo" task failed unexpectedly. [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: System.DllNotFoundException: Unable to load DLL 'api-ms-win-core-sysinfo-l1-2-0.dll': The specified module could not be found. (Exception from HRESULT: 0x8007007E) [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Interop.mincore.GetNativeSystemInfo(SYSTEM_INFO& lpSystemInfo) [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at System.Runtime.InteropServices.RuntimeInformation.get_OSArchitecture() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.DotNet.Build.Tasks.GetTargetMachineInfo.Execute() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.Build.BackEnd.TaskExecutionHost.Microsoft.Build.BackEnd.ITaskExecutionHost.Execute() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] 18:12:03 d:\j\workspace\outerloop_win---9c9e7d59\Tools\publishtest.targets(77,5): error MSB4018: at Microsoft.Build.BackEnd.TaskBuilder.<ExecuteInstantiatedTask>d__26.MoveNext() [d:\j\workspace\outerloop_win---9c9e7d59\src\Common\tests\Common.Tests.csproj] ```
infrastructure
builds failing with gettargetmachineinfo task failed unexpectedly e g d j workspace outerloop win tools publishtest targets error the gettargetmachineinfo task failed unexpectedly d j workspace outerloop win tools publishtest targets error system dllnotfoundexception unable to load dll api ms win core sysinfo dll the specified module could not be found exception from hresult d j workspace outerloop win tools publishtest targets error at interop mincore getnativesysteminfo system info lpsysteminfo d j workspace outerloop win tools publishtest targets error at system runtime interopservices runtimeinformation get osarchitecture d j workspace outerloop win tools publishtest targets error at microsoft dotnet build tasks gettargetmachineinfo execute d j workspace outerloop win tools publishtest targets error at microsoft build backend taskexecutionhost microsoft build backend itaskexecutionhost execute d j workspace outerloop win tools publishtest targets error at microsoft build backend taskbuilder d movenext
1
21,924
14,935,829,292
IssuesEvent
2021-01-25 12:31:59
airyhq/airy
https://api.github.com/repos/airyhq/airy
closed
New Badge Needed: Continuous Integration
docs infrastructure
We should investigate if we merit having a CI Badge, could be done via [Github Actions](https://docs.github.com/en/actions/guides/about-continuous-integration)
1.0
New Badge Needed: Continuous Integration - We should investigate if we merit having a CI Badge, could be done via [Github Actions](https://docs.github.com/en/actions/guides/about-continuous-integration)
infrastructure
new badge needed continuous integration we should investigate if we merit having a ci badge could be done via
1
12,237
9,659,790,743
IssuesEvent
2019-05-20 14:11:35
maidsafe/safe-build-infrastructure
https://api.github.com/repos/maidsafe/safe-build-infrastructure
opened
Add 2nd Static Windows Slave
enhancement infrastructure
We agreed we'd use 2 static slaves to start with. It will take us some time before we can add an internal Windows slave.
1.0
Add 2nd Static Windows Slave - We agreed we'd use 2 static slaves to start with. It will take us some time before we can add an internal Windows slave.
infrastructure
add static windows slave we agreed we d use static slaves to start with it will take us some time before we can add an internal windows slave
1
27,087
21,154,939,535
IssuesEvent
2022-04-07 01:27:24
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
EnabledCallback method of DisplayAttribute
interface/infrastructure question
I cannot find the implementation of the EnabledCallback method name passed as a DisplayAttribute variable in the code. Lots of properties provide a value for this parameter in the DisplayAttributes. I was hoping to utilise this feature or add a similar VisibilityCallback method to not display properties in some Model instances (e.g. when no value provided by the drop down Method). Has this functionality been lost in the upgrade of UI to .Net core?
1.0
EnabledCallback method of DisplayAttribute - I cannot find the implementation of the EnabledCallback method name passed as a DisplayAttribute variable in the code. Lots of properties provide a value for this parameter in the DisplayAttributes. I was hoping to utilise this feature or add a similar VisibilityCallback method to not display properties in some Model instances (e.g. when no value provided by the drop down Method). Has this functionality been lost in the upgrade of UI to .Net core?
infrastructure
enabledcallback method of displayattribute i cannot find the implementation of the enabledcallback method name passed as a displayattribute variable in the code lots of properties provide a value for this parameter in the displayattributes i was hoping to utilise this feature or add a similar visibilitycallback method to not display properties in some model instances e g when no value provided by the drop down method has this functionality been lost in the upgrade of ui to net core
1
49
2,505,050,245
IssuesEvent
2015-01-11 00:13:50
bradfrost/atomic-design
https://api.github.com/repos/bradfrost/atomic-design
opened
Style Text Passages
infrastructure
Need to create styles for text passages, including links, blockquotes, headings, images, media, lists, etc.
1.0
Style Text Passages - Need to create styles for text passages, including links, blockquotes, headings, images, media, lists, etc.
infrastructure
style text passages need to create styles for text passages including links blockquotes headings images media lists etc
1
3,551
4,396,222,910
IssuesEvent
2016-08-10 00:47:30
KhronosGroup/glslang
https://api.github.com/repos/KhronosGroup/glslang
closed
[VS2013 SP5] ShaderLang.h IncludeResult warning prevents compilation when warnings are treated as erorrs
bug Infrastructure
I get the following warnings preventing me from compiling with the latest ShaderLang.h 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): error C2220: warning treated as error - no 'object' file generated 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): warning C4510: 'glslang::TShader::Includer::IncludeResult' : default constructor could not be generated 1> h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(325) : see declaration of 'glslang::TShader::Includer::IncludeResult' 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): warning C4610: struct 'glslang::TShader::Includer::IncludeResult' can never be instantiated - user defined constructor required The problem is that IncludeResult has const members which would need to be initialized by a default constructor. One possible fix would be to do this: ``` struct IncludeResult { // For a successful inclusion, the fully resolved name of the requested // include. For example, in a filesystem-based includer, full resolution // should convert a relative path name into an absolute path name. // For a failed inclusion, this is an empty string. std::string file_name; // The content and byte length of the requested inclusion. The // Includer producing this IncludeResult retains ownership of the // storage. // For a failed inclusion, the file_data // field points to a string containing error details. const char* file_data; size_t file_length; // Include resolver's context. void* user_data; }; ```
1.0
[VS2013 SP5] ShaderLang.h IncludeResult warning prevents compilation when warnings are treated as erorrs - I get the following warnings preventing me from compiling with the latest ShaderLang.h 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): error C2220: warning treated as error - no 'object' file generated 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): warning C4510: 'glslang::TShader::Includer::IncludeResult' : default constructor could not be generated 1> h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(325) : see declaration of 'glslang::TShader::Includer::IncludeResult' 1>h:\dev\source2\main\src\thirdparty\glslang\glslang\public\shaderlang.h(340): warning C4610: struct 'glslang::TShader::Includer::IncludeResult' can never be instantiated - user defined constructor required The problem is that IncludeResult has const members which would need to be initialized by a default constructor. One possible fix would be to do this: ``` struct IncludeResult { // For a successful inclusion, the fully resolved name of the requested // include. For example, in a filesystem-based includer, full resolution // should convert a relative path name into an absolute path name. // For a failed inclusion, this is an empty string. std::string file_name; // The content and byte length of the requested inclusion. The // Includer producing this IncludeResult retains ownership of the // storage. // For a failed inclusion, the file_data // field points to a string containing error details. const char* file_data; size_t file_length; // Include resolver's context. void* user_data; }; ```
infrastructure
shaderlang h includeresult warning prevents compilation when warnings are treated as erorrs i get the following warnings preventing me from compiling with the latest shaderlang h h dev main src thirdparty glslang glslang public shaderlang h error warning treated as error no object file generated h dev main src thirdparty glslang glslang public shaderlang h warning glslang tshader includer includeresult default constructor could not be generated h dev main src thirdparty glslang glslang public shaderlang h see declaration of glslang tshader includer includeresult h dev main src thirdparty glslang glslang public shaderlang h warning struct glslang tshader includer includeresult can never be instantiated user defined constructor required the problem is that includeresult has const members which would need to be initialized by a default constructor one possible fix would be to do this struct includeresult for a successful inclusion the fully resolved name of the requested include for example in a filesystem based includer full resolution should convert a relative path name into an absolute path name for a failed inclusion this is an empty string std string file name the content and byte length of the requested inclusion the includer producing this includeresult retains ownership of the storage for a failed inclusion the file data field points to a string containing error details const char file data size t file length include resolver s context void user data
1
288,656
21,715,948,430
IssuesEvent
2022-05-10 17:55:29
draeger-lab/ModelPolisher
https://api.github.com/repos/draeger-lab/ModelPolisher
opened
write class and method docstrings
documentation
Looking over the javadocs #116 revealed that we have a distinct lack of docstrings in the code base. All classes, public constructors and public methods should carry a docstring.
1.0
write class and method docstrings - Looking over the javadocs #116 revealed that we have a distinct lack of docstrings in the code base. All classes, public constructors and public methods should carry a docstring.
non_infrastructure
write class and method docstrings looking over the javadocs revealed that we have a distinct lack of docstrings in the code base all classes public constructors and public methods should carry a docstring
0
306
3,078,425,253
IssuesEvent
2015-08-21 10:06:31
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
closed
r500. Нелогичное поведение автовыбора настроек соединения.
bug imported Maintainability Priority-Medium
_From [bobrikov](https://code.google.com/u/bobrikov/) on May 06, 2011 12:43:04_ У меня стоит роутер с фключенным UPnP , в настройках соединения ставлю Фаервлл с UPnP. Соединяюсь - он пишеь в логе, что не удалось пробросить и сам включает активный режим. (может роутер глюкнул, не знаю, в стронге было всё ок) Зачем мне активный режим на роутере, если кроме пассива, больше ничего не даст найти ни файла. Да и хабы после поиска выкидывают за подмену ип. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=449_
True
r500. Нелогичное поведение автовыбора настроек соединения. - _From [bobrikov](https://code.google.com/u/bobrikov/) on May 06, 2011 12:43:04_ У меня стоит роутер с фключенным UPnP , в настройках соединения ставлю Фаервлл с UPnP. Соединяюсь - он пишеь в логе, что не удалось пробросить и сам включает активный режим. (может роутер глюкнул, не знаю, в стронге было всё ок) Зачем мне активный режим на роутере, если кроме пассива, больше ничего не даст найти ни файла. Да и хабы после поиска выкидывают за подмену ип. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=449_
non_infrastructure
нелогичное поведение автовыбора настроек соединения from on may у меня стоит роутер с фключенным upnp в настройках соединения ставлю фаервлл с upnp соединяюсь он пишеь в логе что не удалось пробросить и сам включает активный режим может роутер глюкнул не знаю в стронге было всё ок зачем мне активный режим на роутере если кроме пассива больше ничего не даст найти ни файла да и хабы после поиска выкидывают за подмену ип original issue
0
811,749
30,298,915,381
IssuesEvent
2023-07-10 03:10:58
mixcore/mix.core
https://api.github.com/repos/mixcore/mix.core
closed
Portal > limit access and actions to created contents (posts, pages, modules?, database?) that not belonging to the current login user
Priority:2 v2.x
**Describe the bug** At the moment if user A created Post "ABC", then user B still able to see user A's posts and edit those posts. **Suggestion**: Implement default roles/permissions Summary of default Roles - Super Admin – somebody with access to the site network administration features and all other features. See the Create a Network article. - Administrator (slug: ‘administrator’) – somebody who has access to all the administration features within a single site. - Editor (slug: ‘editor’) – somebody who can publish and manage posts including the posts of other users. - Author (slug: ‘author’) – somebody who can publish and manage their own posts. - Contributor (slug: ‘contributor’) – somebody who can write and manage their own posts but cannot publish them. - Subscriber (slug: ‘subscriber’) – somebody who can only manage their profile.
1.0
Portal > limit access and actions to created contents (posts, pages, modules?, database?) that not belonging to the current login user - **Describe the bug** At the moment if user A created Post "ABC", then user B still able to see user A's posts and edit those posts. **Suggestion**: Implement default roles/permissions Summary of default Roles - Super Admin – somebody with access to the site network administration features and all other features. See the Create a Network article. - Administrator (slug: ‘administrator’) – somebody who has access to all the administration features within a single site. - Editor (slug: ‘editor’) – somebody who can publish and manage posts including the posts of other users. - Author (slug: ‘author’) – somebody who can publish and manage their own posts. - Contributor (slug: ‘contributor’) – somebody who can write and manage their own posts but cannot publish them. - Subscriber (slug: ‘subscriber’) – somebody who can only manage their profile.
non_infrastructure
portal limit access and actions to created contents posts pages modules database that not belonging to the current login user describe the bug at the moment if user a created post abc then user b still able to see user a s posts and edit those posts suggestion implement default roles permissions summary of default roles super admin – somebody with access to the site network administration features and all other features see the create a network article administrator slug ‘administrator’ – somebody who has access to all the administration features within a single site editor slug ‘editor’ – somebody who can publish and manage posts including the posts of other users author slug ‘author’ – somebody who can publish and manage their own posts contributor slug ‘contributor’ – somebody who can write and manage their own posts but cannot publish them subscriber slug ‘subscriber’ – somebody who can only manage their profile
0
22,267
4,786,801,195
IssuesEvent
2016-10-29 16:48:48
ngageoint/scale
https://api.github.com/repos/ngageoint/scale
opened
Scale logo 404s in DCOS deployment of docs
bug deployment documentation
We need to fix the broken link to the Scale logo when the docs are deployed in DCOS: /service/scale/docs/_static/scale3-transparent-128.png returns a 404
1.0
Scale logo 404s in DCOS deployment of docs - We need to fix the broken link to the Scale logo when the docs are deployed in DCOS: /service/scale/docs/_static/scale3-transparent-128.png returns a 404
non_infrastructure
scale logo in dcos deployment of docs we need to fix the broken link to the scale logo when the docs are deployed in dcos service scale docs static transparent png returns a
0
32,266
26,583,175,104
IssuesEvent
2023-01-22 18:05:25
iree-org/iree
https://api.github.com/repos/iree-org/iree
closed
Cannot consume iree runtime with cpm or cmake FetchContent (missing install directives for libraries ?)
bug 🐞 infrastructure support
### What happened? I would like to add iree as a dependency of a project using cmake FetchContent. It makes consuming/depending on libraries really easy, and would help folks integrate iree in their projects. Here is a good [write up on it](https://medium.com/analytics-vidhya/c-dependency-management-with-cmakes-fetchcontent-4ceca4693a5d). There is also a cmake project that makes consumption (aka adding a depenrency on an external project) even easier, called [CPM](https://github.com/cpm-cmake/CPM.cmake) (for cmake package manager), which is actually what we use. The error at cmake configure time reads like: ``` xxx which requires target "iree_base_base" that is not in any export set. ``` ### Steps to reproduce your issue 1. Take one example from https://medium.com/analytics-vidhya/c-dependency-management-with-cmakes-fetchcontent-4ceca4693a5d and replace a third-party library like DocTest with iree. ### What component(s) does this issue relate to? Runtime ### Version information I'm using the main/master branch as of December 3rd, 2022, so I think around dc15a278e838bb95e39560c24fdea3773d338a16 ### Additional context It looks like a wrapper is used for declaring libraries, to keep the code tidy -> build_tools/cmake/iree_cc_library.cmake Using this example / https://github.com/maxmind/libmaxminddb/blob/main/CMakeLists.txt#L91 My understanding is that 2 directives are required. I am not a cmake expert ... ``` install(TARGETS maxminddb EXPORT maxminddb) # This is required to work with FetchContent install(EXPORT maxminddb FILE maxminddb-config.cmake NAMESPACE maxminddb:: DESTINATION ${CMAKE_INSTALL_LIBDIR}/cmake/maxminddb) ``` I tried to use the source tarball from September but I ran into other problems, that seems to be resolved in master.
1.0
Cannot consume iree runtime with cpm or cmake FetchContent (missing install directives for libraries ?) - ### What happened? I would like to add iree as a dependency of a project using cmake FetchContent. It makes consuming/depending on libraries really easy, and would help folks integrate iree in their projects. Here is a good [write up on it](https://medium.com/analytics-vidhya/c-dependency-management-with-cmakes-fetchcontent-4ceca4693a5d). There is also a cmake project that makes consumption (aka adding a depenrency on an external project) even easier, called [CPM](https://github.com/cpm-cmake/CPM.cmake) (for cmake package manager), which is actually what we use. The error at cmake configure time reads like: ``` xxx which requires target "iree_base_base" that is not in any export set. ``` ### Steps to reproduce your issue 1. Take one example from https://medium.com/analytics-vidhya/c-dependency-management-with-cmakes-fetchcontent-4ceca4693a5d and replace a third-party library like DocTest with iree. ### What component(s) does this issue relate to? Runtime ### Version information I'm using the main/master branch as of December 3rd, 2022, so I think around dc15a278e838bb95e39560c24fdea3773d338a16 ### Additional context It looks like a wrapper is used for declaring libraries, to keep the code tidy -> build_tools/cmake/iree_cc_library.cmake Using this example / https://github.com/maxmind/libmaxminddb/blob/main/CMakeLists.txt#L91 My understanding is that 2 directives are required. I am not a cmake expert ... ``` install(TARGETS maxminddb EXPORT maxminddb) # This is required to work with FetchContent install(EXPORT maxminddb FILE maxminddb-config.cmake NAMESPACE maxminddb:: DESTINATION ${CMAKE_INSTALL_LIBDIR}/cmake/maxminddb) ``` I tried to use the source tarball from September but I ran into other problems, that seems to be resolved in master.
infrastructure
cannot consume iree runtime with cpm or cmake fetchcontent missing install directives for libraries what happened i would like to add iree as a dependency of a project using cmake fetchcontent it makes consuming depending on libraries really easy and would help folks integrate iree in their projects here is a good there is also a cmake project that makes consumption aka adding a depenrency on an external project even easier called for cmake package manager which is actually what we use the error at cmake configure time reads like xxx which requires target iree base base that is not in any export set steps to reproduce your issue take one example from and replace a third party library like doctest with iree what component s does this issue relate to runtime version information i m using the main master branch as of december so i think around additional context it looks like a wrapper is used for declaring libraries to keep the code tidy build tools cmake iree cc library cmake using this example my understanding is that directives are required i am not a cmake expert install targets maxminddb export maxminddb this is required to work with fetchcontent install export maxminddb file maxminddb config cmake namespace maxminddb destination cmake install libdir cmake maxminddb i tried to use the source tarball from september but i ran into other problems that seems to be resolved in master
1
124,031
12,223,852,810
IssuesEvent
2020-05-02 19:31:59
anitab-org/mentorship-flutter
https://api.github.com/repos/anitab-org/mentorship-flutter
closed
Update README with how to setup and run flutter instructions.
First Timers Only documentation
## Description As a developer,, I need to make sure contributors can setup and run flutter on their local machine so they can start contributing to the project, ## Mocks N/A ## Acceptance Criteria - [ ] Update README.md with setup instructions ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 1 hour
1.0
Update README with how to setup and run flutter instructions. - ## Description As a developer,, I need to make sure contributors can setup and run flutter on their local machine so they can start contributing to the project, ## Mocks N/A ## Acceptance Criteria - [ ] Update README.md with setup instructions ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 1 hour
non_infrastructure
update readme with how to setup and run flutter instructions description as a developer i need to make sure contributors can setup and run flutter on their local machine so they can start contributing to the project mocks n a acceptance criteria update readme md with setup instructions definition of done all of the required items are completed approval by mentor estimation hour
0
13,523
10,313,075,419
IssuesEvent
2019-08-29 21:32:54
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
opened
celotool terraform commands should always init before
infrastructure
### Expected Behavior `terraform init` always run before another celotool command that is not a part of the deployment process (like refresh, NOT apply etc) ### Current Behavior Some commands do not run `terraform init` before. This is bad for cases where a backend switch is required
1.0
celotool terraform commands should always init before - ### Expected Behavior `terraform init` always run before another celotool command that is not a part of the deployment process (like refresh, NOT apply etc) ### Current Behavior Some commands do not run `terraform init` before. This is bad for cases where a backend switch is required
infrastructure
celotool terraform commands should always init before expected behavior terraform init always run before another celotool command that is not a part of the deployment process like refresh not apply etc current behavior some commands do not run terraform init before this is bad for cases where a backend switch is required
1
167,677
20,726,277,937
IssuesEvent
2022-03-14 02:32:00
kapseliboi/webuild
https://api.github.com/repos/kapseliboi/webuild
opened
WS-2020-0342 (High) detected in is-my-json-valid-2.17.2.tgz
security vulnerability
## WS-2020-0342 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.2.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.2.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - webuild-events-1.4.39.tgz (Root Library) - ical-0.5.0.tgz - request-2.68.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.17.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format. <p>Publish Date: 2020-06-27 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb>WS-2020-0342</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-27</p> <p>Fix Resolution: is-my-json-valid - 2.20.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0342 (High) detected in is-my-json-valid-2.17.2.tgz - ## WS-2020-0342 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.2.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.2.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - webuild-events-1.4.39.tgz (Root Library) - ical-0.5.0.tgz - request-2.68.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.17.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format. <p>Publish Date: 2020-06-27 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb>WS-2020-0342</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-27</p> <p>Fix Resolution: is-my-json-valid - 2.20.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
ws high detected in is my json valid tgz ws high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules is my json valid package json dependency hierarchy webuild events tgz root library ical tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in base branch master vulnerability details regular expression denial of service redos vulnerability was found in is my json valid before via the style format publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid step up your open source security game with whitesource
0
142,964
5,487,233,673
IssuesEvent
2017-03-14 03:30:04
kamal1978/LTFHC
https://api.github.com/repos/kamal1978/LTFHC
opened
Disable internet at start of application
app enhancement low priority
Per a conversation with Kamal: Disabling internet when starting the application will help to conserve power. Internet should become active when upload button is pressed.
1.0
Disable internet at start of application - Per a conversation with Kamal: Disabling internet when starting the application will help to conserve power. Internet should become active when upload button is pressed.
non_infrastructure
disable internet at start of application per a conversation with kamal disabling internet when starting the application will help to conserve power internet should become active when upload button is pressed
0
12,480
9,799,534,986
IssuesEvent
2019-06-11 14:35:40
coq/coq
https://api.github.com/repos/coq/coq
closed
big-memory CI worker for fiat-crypto
kind: infrastructure
I'd like to enable the more memory-intensive targets of fiat-crypto on Coq's CI, as they have recently been broken by an update to Coq (see #10352). Peak memory usage per file seems to be 2 385 020 ko on master, a bit more on 8.9 (2 526 224 ko) and 8.8 (2 799 940 ko). The relevant change to https://github.com/coq/coq/blob/45306c6c9c433b86406d041f58aafb7cf3a3ff82/dev/ci/ci-fiat-crypto.sh#L13-L18 is to change `print-nobigmem nobigmem` to `coq` cc @ejgallego What needs to be done on my side?
1.0
big-memory CI worker for fiat-crypto - I'd like to enable the more memory-intensive targets of fiat-crypto on Coq's CI, as they have recently been broken by an update to Coq (see #10352). Peak memory usage per file seems to be 2 385 020 ko on master, a bit more on 8.9 (2 526 224 ko) and 8.8 (2 799 940 ko). The relevant change to https://github.com/coq/coq/blob/45306c6c9c433b86406d041f58aafb7cf3a3ff82/dev/ci/ci-fiat-crypto.sh#L13-L18 is to change `print-nobigmem nobigmem` to `coq` cc @ejgallego What needs to be done on my side?
infrastructure
big memory ci worker for fiat crypto i d like to enable the more memory intensive targets of fiat crypto on coq s ci as they have recently been broken by an update to coq see peak memory usage per file seems to be ko on master a bit more on ko and ko the relevant change to is to change print nobigmem nobigmem to coq cc ejgallego what needs to be done on my side
1
430,014
12,439,878,315
IssuesEvent
2020-05-26 10:55:21
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
opened
Disable SNMP Topology when not available
low-priority bug user interface
In case LLDP is not available for a given host, the topology badge and company icon must be removed ![Screen Shot 2020-05-26 at 12 54 30](https://user-images.githubusercontent.com/4493366/82892660-24b0fb80-9f50-11ea-8fc0-03162ef810ca.png)
1.0
Disable SNMP Topology when not available - In case LLDP is not available for a given host, the topology badge and company icon must be removed ![Screen Shot 2020-05-26 at 12 54 30](https://user-images.githubusercontent.com/4493366/82892660-24b0fb80-9f50-11ea-8fc0-03162ef810ca.png)
non_infrastructure
disable snmp topology when not available in case lldp is not available for a given host the topology badge and company icon must be removed
0
158,911
12,438,677,647
IssuesEvent
2020-05-26 08:50:15
mautic/mautic
https://api.github.com/repos/mautic/mautic
closed
Send webhook doesn't properly encode email address with '+' symbol
Bug Ready To Test
**Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.** What type of report is this: | Q | A | ---| --- | Bug report? | Y | Feature request? | N | Enhancement? | N ## Description: When a webhook action is sent from a campaign that includes the contact's e-mail address token, if the e-mail address contains a '+ symbol, the symbol comes across as a space instead. ## If a bug: | Q | A | --- | --- | Mautic version | 2.13.1 | PHP version | 5.6.33 ### Steps to reproduce: 1. Create a contact with an e-mail address that includes a '+' symbol. ie. 'me+1@home.net' 2. Create a segment and add this contact so we can include in a new campaign. 3. Create a campaign for that segment that includes a "Send a webhook" action, method set to GET. 4. In the data, include the contact's e-mail address using the token {leadfield=email} 5. On the server side, check the e-mail address that is received, 'me 1@home.net' does not match the contact's e-mail address, 'me+1@home.net'. ### Log errors: N/A _Please check for related errors in the latest log file in [mautic root]/app/log/ and/or the web server's logs and post them here. Be sure to remove sensitive information if applicable._
1.0
Send webhook doesn't properly encode email address with '+' symbol - **Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.** What type of report is this: | Q | A | ---| --- | Bug report? | Y | Feature request? | N | Enhancement? | N ## Description: When a webhook action is sent from a campaign that includes the contact's e-mail address token, if the e-mail address contains a '+ symbol, the symbol comes across as a space instead. ## If a bug: | Q | A | --- | --- | Mautic version | 2.13.1 | PHP version | 5.6.33 ### Steps to reproduce: 1. Create a contact with an e-mail address that includes a '+' symbol. ie. 'me+1@home.net' 2. Create a segment and add this contact so we can include in a new campaign. 3. Create a campaign for that segment that includes a "Send a webhook" action, method set to GET. 4. In the data, include the contact's e-mail address using the token {leadfield=email} 5. On the server side, check the e-mail address that is received, 'me 1@home.net' does not match the contact's e-mail address, 'me+1@home.net'. ### Log errors: N/A _Please check for related errors in the latest log file in [mautic root]/app/log/ and/or the web server's logs and post them here. Be sure to remove sensitive information if applicable._
non_infrastructure
send webhook doesn t properly encode email address with symbol please do not report security vulnerabilities here send them to security mautic com instead what type of report is this q a bug report y feature request n enhancement n description when a webhook action is sent from a campaign that includes the contact s e mail address token if the e mail address contains a symbol the symbol comes across as a space instead if a bug q a mautic version php version steps to reproduce create a contact with an e mail address that includes a symbol ie me home net create a segment and add this contact so we can include in a new campaign create a campaign for that segment that includes a send a webhook action method set to get in the data include the contact s e mail address using the token leadfield email on the server side check the e mail address that is received me home net does not match the contact s e mail address me home net log errors n a please check for related errors in the latest log file in app log and or the web server s logs and post them here be sure to remove sensitive information if applicable
0
7,588
7,005,113,599
IssuesEvent
2017-12-19 00:04:25
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Apostrophes should be allowed in bugdown tests
area: testing-infrastructure help wanted
# Description When a fixture in `zerver/fixtures/markdown_test_cases.json` contains an apostrophe (`'`), the result produced by the frontend (Marked) and the backend (Bugdown) don't seem to match, since running the test with the `bugdown_matches_marked` flag as true produces an `AssertionError`. # Steps to reproduce 1. Add the following key to `zerver/fixtures/markdown_test_cases.json`, inside the `regular_tests` array: ``` { "name": "test", "input": "it's lunch time", "expected_output": "<p>it's lunch time</p>", "bugdown_matches_marked": true } ``` 2. Run `./tools/test-js-with-node`. 3. The following error appears: ``` assert.js:85 throw new assert.AssertionError({ ^ AssertionError: '<p>it\'s lunch time</p>' == '<p>it&#39;s lunch time</p>' at /srv/zulip/frontend_tests/node_tests/echo.js:159:14 at Array.forEach (native) at test_marked_shared (/srv/zulip/frontend_tests/node_tests/echo.js:153:9) at Object.<anonymous> (/srv/zulip/frontend_tests/node_tests/echo.js:164:2) at Module._compile (module.js:556:32) at Object.Module._extensions..js (module.js:565:10) at Module.load (module.js:473:32) at tryModuleLoad (module.js:432:12) at Function.Module._load (module.js:424:3) at Module.require (module.js:483:17) at require (internal/module.js:20:19) at /srv/zulip/frontend_tests/zjsunit/index.js:52:5 at Array.forEach (native) at Object.<anonymous> (/srv/zulip/frontend_tests/zjsunit/index.js:46:7) at Module._compile (module.js:556:32) at Object.Module._extensions..js (module.js:565:10) at Module.load (module.js:473:32) at tryModuleLoad (module.js:432:12) at Function.Module._load (module.js:424:3) at Module.runMain (module.js:590:10) at run (bootstrap_node.js:394:7) at startup (bootstrap_node.js:149:9) at bootstrap_node.js:509:3 FAIL - Test(s) failed ``` # Additional information At first sight, it could look like this is caused due to the frontend escaping apostrophes as `&#39;`, while the backend doesn't do it. However, if you try to type a message with an apostrophe just in local echo (with the server disconnected), the HTML directly contains the character, with no escaping whatsoever. This is probably related to the `escape` functions defined either in `static/js/echo.js` or `static/js/third/marked/lib/marked.js`. More information on this issue and how it was discovered can be found in [this](https://chat.zulip.org/#narrow/stream/test.20suites/topic/aphostrophes.20in.20bugdown.20tests) Zulip conversation.
1.0
Apostrophes should be allowed in bugdown tests - # Description When a fixture in `zerver/fixtures/markdown_test_cases.json` contains an apostrophe (`'`), the result produced by the frontend (Marked) and the backend (Bugdown) don't seem to match, since running the test with the `bugdown_matches_marked` flag as true produces an `AssertionError`. # Steps to reproduce 1. Add the following key to `zerver/fixtures/markdown_test_cases.json`, inside the `regular_tests` array: ``` { "name": "test", "input": "it's lunch time", "expected_output": "<p>it's lunch time</p>", "bugdown_matches_marked": true } ``` 2. Run `./tools/test-js-with-node`. 3. The following error appears: ``` assert.js:85 throw new assert.AssertionError({ ^ AssertionError: '<p>it\'s lunch time</p>' == '<p>it&#39;s lunch time</p>' at /srv/zulip/frontend_tests/node_tests/echo.js:159:14 at Array.forEach (native) at test_marked_shared (/srv/zulip/frontend_tests/node_tests/echo.js:153:9) at Object.<anonymous> (/srv/zulip/frontend_tests/node_tests/echo.js:164:2) at Module._compile (module.js:556:32) at Object.Module._extensions..js (module.js:565:10) at Module.load (module.js:473:32) at tryModuleLoad (module.js:432:12) at Function.Module._load (module.js:424:3) at Module.require (module.js:483:17) at require (internal/module.js:20:19) at /srv/zulip/frontend_tests/zjsunit/index.js:52:5 at Array.forEach (native) at Object.<anonymous> (/srv/zulip/frontend_tests/zjsunit/index.js:46:7) at Module._compile (module.js:556:32) at Object.Module._extensions..js (module.js:565:10) at Module.load (module.js:473:32) at tryModuleLoad (module.js:432:12) at Function.Module._load (module.js:424:3) at Module.runMain (module.js:590:10) at run (bootstrap_node.js:394:7) at startup (bootstrap_node.js:149:9) at bootstrap_node.js:509:3 FAIL - Test(s) failed ``` # Additional information At first sight, it could look like this is caused due to the frontend escaping apostrophes as `&#39;`, while the backend doesn't do it. However, if you try to type a message with an apostrophe just in local echo (with the server disconnected), the HTML directly contains the character, with no escaping whatsoever. This is probably related to the `escape` functions defined either in `static/js/echo.js` or `static/js/third/marked/lib/marked.js`. More information on this issue and how it was discovered can be found in [this](https://chat.zulip.org/#narrow/stream/test.20suites/topic/aphostrophes.20in.20bugdown.20tests) Zulip conversation.
infrastructure
apostrophes should be allowed in bugdown tests description when a fixture in zerver fixtures markdown test cases json contains an apostrophe the result produced by the frontend marked and the backend bugdown don t seem to match since running the test with the bugdown matches marked flag as true produces an assertionerror steps to reproduce add the following key to zerver fixtures markdown test cases json inside the regular tests array name test input it s lunch time expected output it s lunch time bugdown matches marked true run tools test js with node the following error appears assert js throw new assert assertionerror assertionerror it s lunch time it s lunch time at srv zulip frontend tests node tests echo js at array foreach native at test marked shared srv zulip frontend tests node tests echo js at object srv zulip frontend tests node tests echo js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js at function module load module js at module require module js at require internal module js at srv zulip frontend tests zjsunit index js at array foreach native at object srv zulip frontend tests zjsunit index js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js at function module load module js at module runmain module js at run bootstrap node js at startup bootstrap node js at bootstrap node js fail test s failed additional information at first sight it could look like this is caused due to the frontend escaping apostrophes as while the backend doesn t do it however if you try to type a message with an apostrophe just in local echo with the server disconnected the html directly contains the character with no escaping whatsoever this is probably related to the escape functions defined either in static js echo js or static js third marked lib marked js more information on this issue and how it was discovered can be found in zulip conversation
1
31,255
25,484,255,852
IssuesEvent
2022-11-26 06:34:42
UBCSailbot/sailbot_workspace
https://api.github.com/repos/UBCSailbot/sailbot_workspace
opened
Hacks
infrastructure
### Purpose <!-- What is the purpose of this issue? What does this issue hope to achieve? --> [Hopefully] temporary workarounds for various infrastructure related items ### Hacks - [ ] Downgrade setuptools: https://answers.ros.org/question/396439/setuptoolsdeprecationwarning-setuppy-install-is-deprecated-use-build-and-pip-and-other-standards-based-tools/?answer=400052#post-id-400052
1.0
Hacks - ### Purpose <!-- What is the purpose of this issue? What does this issue hope to achieve? --> [Hopefully] temporary workarounds for various infrastructure related items ### Hacks - [ ] Downgrade setuptools: https://answers.ros.org/question/396439/setuptoolsdeprecationwarning-setuppy-install-is-deprecated-use-build-and-pip-and-other-standards-based-tools/?answer=400052#post-id-400052
infrastructure
hacks purpose temporary workarounds for various infrastructure related items hacks downgrade setuptools
1
136,831
5,289,146,421
IssuesEvent
2017-02-08 16:43:05
rd-alliance/metadata-catalog-dev
https://api.github.com/repos/rd-alliance/metadata-catalog-dev
opened
Display maturity of scheme
GUI low priority
Display automatically calculated maturity rating. The algorithm could take into account the status of the maintainer, the version history, number of known users, and so on.
1.0
Display maturity of scheme - Display automatically calculated maturity rating. The algorithm could take into account the status of the maintainer, the version history, number of known users, and so on.
non_infrastructure
display maturity of scheme display automatically calculated maturity rating the algorithm could take into account the status of the maintainer the version history number of known users and so on
0
65,161
7,859,264,316
IssuesEvent
2018-06-21 16:04:37
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
Update installation documentation to redirect to sct_docker repos
card:DESIGN_DISCUSSION installation priority:HIGH
### Description Now that we have a working solution for Windows users via Docker, maybe we want to replace the VM solution with the Docker solution? If so, then we should update the README on SCT repos. In the future, it will be easier to only maintain the Docker repos (as opposed to Docker+VM).
1.0
Update installation documentation to redirect to sct_docker repos - ### Description Now that we have a working solution for Windows users via Docker, maybe we want to replace the VM solution with the Docker solution? If so, then we should update the README on SCT repos. In the future, it will be easier to only maintain the Docker repos (as opposed to Docker+VM).
non_infrastructure
update installation documentation to redirect to sct docker repos description now that we have a working solution for windows users via docker maybe we want to replace the vm solution with the docker solution if so then we should update the readme on sct repos in the future it will be easier to only maintain the docker repos as opposed to docker vm
0
727,880
25,049,550,155
IssuesEvent
2022-11-05 18:01:46
projectsend/projectsend
https://api.github.com/repos/projectsend/projectsend
closed
Security Vulnerability
Bug Security High Priority
Hi @ignacionelson , I have tried to contact you through contact@projectsend.org but unfortunately no one is answering to me. You have critical security vulnerabilities in your code. Could you please tell me, how you plan to approach this issue or take any other action? Otherwise I have to publish the security problem here on Gitlab. Best Regards mschop
1.0
Security Vulnerability - Hi @ignacionelson , I have tried to contact you through contact@projectsend.org but unfortunately no one is answering to me. You have critical security vulnerabilities in your code. Could you please tell me, how you plan to approach this issue or take any other action? Otherwise I have to publish the security problem here on Gitlab. Best Regards mschop
non_infrastructure
security vulnerability hi ignacionelson i have tried to contact you through contact projectsend org but unfortunately no one is answering to me you have critical security vulnerabilities in your code could you please tell me how you plan to approach this issue or take any other action otherwise i have to publish the security problem here on gitlab best regards mschop
0
257,900
8,148,392,973
IssuesEvent
2018-08-22 05:33:26
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
closed
The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2).
Bug C: Grid F: PDF Export Kendo1 Needs QA Priority 2 S: Wrappers (ASP.NET Core) SEV: Medium
### Bug report The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2). ![image](https://user-images.githubusercontent.com/11087921/31273267-89028dc6-aa96-11e7-9b3e-cf76b9fda387.png) ### Reproduction of the problem The issue can be reproduced in our demo when the filtering is on and the paperSize is set to auto in order to show a bigger part of the Grid. ### Environment * **Browser:** [all]
1.0
The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2). - ### Bug report The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2). ![image](https://user-images.githubusercontent.com/11087921/31273267-89028dc6-aa96-11e7-9b3e-cf76b9fda387.png) ### Reproduction of the problem The issue can be reproduced in our demo when the filtering is on and the paperSize is set to auto in order to show a bigger part of the Grid. ### Environment * **Browser:** [all]
non_infrastructure
the grid pdf export is not showing the icons and it is rendering the wrong header in mvc core bootstrap and bug report the grid pdf export is not showing the icons and it is rendering the wrong header in mvc core bootstrap and reproduction of the problem the issue can be reproduced in our demo when the filtering is on and the papersize is set to auto in order to show a bigger part of the grid environment browser
0
552,582
16,244,082,403
IssuesEvent
2021-05-07 12:58:51
google/ground-platform
https://api.github.com/repos/google/ground-platform
closed
[ACLs] Hide layers in add point dropdown if not allowed
priority: p2 type: feature request web
Hide layers in "add point" flow if not allowed, hide pin if no layers are allowed. Owners and managers may always add points. Contributors may only add points if layer is flagged "contributors may add points". Viewers can never add points. A new method can be added to `AuthManager`, `boolean canAddPoint(layer: Layer)`. @DaoyuT Wdyt?
1.0
[ACLs] Hide layers in add point dropdown if not allowed - Hide layers in "add point" flow if not allowed, hide pin if no layers are allowed. Owners and managers may always add points. Contributors may only add points if layer is flagged "contributors may add points". Viewers can never add points. A new method can be added to `AuthManager`, `boolean canAddPoint(layer: Layer)`. @DaoyuT Wdyt?
non_infrastructure
hide layers in add point dropdown if not allowed hide layers in add point flow if not allowed hide pin if no layers are allowed owners and managers may always add points contributors may only add points if layer is flagged contributors may add points viewers can never add points a new method can be added to authmanager boolean canaddpoint layer layer daoyut wdyt
0
292,978
25,255,460,069
IssuesEvent
2022-11-15 17:40:43
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
Missing gcr.io/authenticated-image-pulling/windows-nanoserver:v3 image for E2E tests
kind/failing-test
### Which jobs are failing? capz-windows-containerd-master ### Which tests are failing? ``[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull from private registry with secret [NodeConformance]`` ### Since when has it been failing? Since https://github.com/kubernetes/kubernetes/pull/113900 merged. ### Testgrid link https://testgrid.k8s.io/sig-windows-signal#capz-windows-containerd-master ### Reason for failure (if possible) A recent PR [1] updated the image versions we use for E2E tests. However, the windows-nanoserver image is meant to be in a private authenticated registry: ``gcr.io/authenticated-image-pulling/windows-nanoserver``, which requires credentials to pull images from it. This image is required by the ``[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull from private registry with secret [NodeConformance]`` test for Windows. The v3 image does not exist, there's no automatic promotion process for that registry. Previously, it was built and pushed manually. Because of this, the https://testgrid.k8s.io/sig-windows-signal#capz-windows-containerd-master jobs have started to fail. [1] https://github.com/kubernetes/kubernetes/pull/113900 ### Anything else we need to know? Related PR: https://github.com/kubernetes/kubernetes/pull/113923 The issue can be resolved if the image is built and pushed to the ``gcr.io/authenticated-image-pulling/`` registry. This can be done by running: ``` cd kubernetes/test/images make REGISTRY=gcr.io/authenticated-image-pulling WHAT=windows-nanoserver all-push ``` ### Relevant SIG(s) /sig windows /sig testing
1.0
Missing gcr.io/authenticated-image-pulling/windows-nanoserver:v3 image for E2E tests - ### Which jobs are failing? capz-windows-containerd-master ### Which tests are failing? ``[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull from private registry with secret [NodeConformance]`` ### Since when has it been failing? Since https://github.com/kubernetes/kubernetes/pull/113900 merged. ### Testgrid link https://testgrid.k8s.io/sig-windows-signal#capz-windows-containerd-master ### Reason for failure (if possible) A recent PR [1] updated the image versions we use for E2E tests. However, the windows-nanoserver image is meant to be in a private authenticated registry: ``gcr.io/authenticated-image-pulling/windows-nanoserver``, which requires credentials to pull images from it. This image is required by the ``[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull from private registry with secret [NodeConformance]`` test for Windows. The v3 image does not exist, there's no automatic promotion process for that registry. Previously, it was built and pushed manually. Because of this, the https://testgrid.k8s.io/sig-windows-signal#capz-windows-containerd-master jobs have started to fail. [1] https://github.com/kubernetes/kubernetes/pull/113900 ### Anything else we need to know? Related PR: https://github.com/kubernetes/kubernetes/pull/113923 The issue can be resolved if the image is built and pushed to the ``gcr.io/authenticated-image-pulling/`` registry. This can be done by running: ``` cd kubernetes/test/images make REGISTRY=gcr.io/authenticated-image-pulling WHAT=windows-nanoserver all-push ``` ### Relevant SIG(s) /sig windows /sig testing
non_infrastructure
missing gcr io authenticated image pulling windows nanoserver image for tests which jobs are failing capz windows containerd master which tests are failing container runtime blackbox test when running a container with a new image should be able to pull from private registry with secret since when has it been failing since merged testgrid link reason for failure if possible a recent pr updated the image versions we use for tests however the windows nanoserver image is meant to be in a private authenticated registry gcr io authenticated image pulling windows nanoserver which requires credentials to pull images from it this image is required by the container runtime blackbox test when running a container with a new image should be able to pull from private registry with secret test for windows the image does not exist there s no automatic promotion process for that registry previously it was built and pushed manually because of this the jobs have started to fail anything else we need to know related pr the issue can be resolved if the image is built and pushed to the gcr io authenticated image pulling registry this can be done by running cd kubernetes test images make registry gcr io authenticated image pulling what windows nanoserver all push relevant sig s sig windows sig testing
0
797,312
28,143,570,966
IssuesEvent
2023-04-02 08:04:26
prodigy-sub/snitch
https://api.github.com/repos/prodigy-sub/snitch
closed
[feature] each news website has a different HTML structure
feature priority-do-it-now
**Is your feature request related to a problem? Please describe.** each news websites have different HTML structure **Describe the solution you'd like** narrow the scope of scrapping (only articles which are provided from NAVER news) **Describe alternatives you've considered** - implement a scrapper for each website - use a AI-based scrapper **TODOs** - [ ] implement a scrapper for NAVER news **Additional context** Add any other context or screenshots about the feature request here.
1.0
[feature] each news website has a different HTML structure - **Is your feature request related to a problem? Please describe.** each news websites have different HTML structure **Describe the solution you'd like** narrow the scope of scrapping (only articles which are provided from NAVER news) **Describe alternatives you've considered** - implement a scrapper for each website - use a AI-based scrapper **TODOs** - [ ] implement a scrapper for NAVER news **Additional context** Add any other context or screenshots about the feature request here.
non_infrastructure
each news website has a different html structure is your feature request related to a problem please describe each news websites have different html structure describe the solution you d like narrow the scope of scrapping only articles which are provided from naver news describe alternatives you ve considered implement a scrapper for each website use a ai based scrapper todos implement a scrapper for naver news additional context add any other context or screenshots about the feature request here
0
23,094
15,818,812,682
IssuesEvent
2021-04-05 16:34:18
OregonDigital/OD2
https://api.github.com/repos/OregonDigital/OD2
closed
volume for migration
Infrastructure
Following up on conversation on Slack: Migration folks would like to have a volume that is mounted on both the OD1 and OD2 worker containers, for the purposes of exporting/migrating assets from OD1 to OD2; suggested mount location /migration and configuration settings (eg, "disable compression and de-dup and a few of the other knobs available for configuring it for speed over storage capacity") all sound good.
1.0
volume for migration - Following up on conversation on Slack: Migration folks would like to have a volume that is mounted on both the OD1 and OD2 worker containers, for the purposes of exporting/migrating assets from OD1 to OD2; suggested mount location /migration and configuration settings (eg, "disable compression and de-dup and a few of the other knobs available for configuring it for speed over storage capacity") all sound good.
infrastructure
volume for migration following up on conversation on slack migration folks would like to have a volume that is mounted on both the and worker containers for the purposes of exporting migrating assets from to suggested mount location migration and configuration settings eg disable compression and de dup and a few of the other knobs available for configuring it for speed over storage capacity all sound good
1
60,957
14,596,423,518
IssuesEvent
2020-12-20 15:47:36
billmcchesney1/superagent
https://api.github.com/repos/billmcchesney1/superagent
opened
CVE-2019-1010266 (Medium) detected in multiple libraries
security vulnerability
## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.1.0.tgz</b>, <b>lodash-3.2.0.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.1.0.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.1.0.tgz">https://registry.npmjs.org/lodash/-/lodash-2.1.0.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/file-utils/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - firefox-profile-0.2.7.tgz - archiver-0.7.1.tgz - file-utils-0.1.5.tgz - :x: **lodash-2.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.2.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.2.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.2.0.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/istanbul-middleware/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - istanbul-middleware-0.2.2.tgz - archiver-0.14.4.tgz - :x: **lodash-3.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/archiver/node_modules/lodash/package.json,superagent/node_modules/findup-sync/node_modules/lodash/package.json,superagent/node_modules/wd/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - wd-0.3.11.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/zuul/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/superagent/commit/77fefdaffd4ef3cef2e5b252e165b5f40fae61d5">77fefdaffd4ef3cef2e5b252e165b5f40fae61d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.1.0","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;firefox-profile:0.2.7;archiver:0.7.1;file-utils:0.1.5;lodash:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.2.0","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;istanbul-middleware:0.2.2;archiver:0.14.4;lodash:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;wd:0.3.11;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-1010266 (Medium) detected in multiple libraries - ## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.1.0.tgz</b>, <b>lodash-3.2.0.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.1.0.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.1.0.tgz">https://registry.npmjs.org/lodash/-/lodash-2.1.0.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/file-utils/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - firefox-profile-0.2.7.tgz - archiver-0.7.1.tgz - file-utils-0.1.5.tgz - :x: **lodash-2.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.2.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.2.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.2.0.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/istanbul-middleware/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - istanbul-middleware-0.2.2.tgz - archiver-0.14.4.tgz - :x: **lodash-3.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/archiver/node_modules/lodash/package.json,superagent/node_modules/findup-sync/node_modules/lodash/package.json,superagent/node_modules/wd/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - wd-0.3.11.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: superagent/package.json</p> <p>Path to vulnerable library: superagent/node_modules/zuul/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zuul-3.12.0.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/superagent/commit/77fefdaffd4ef3cef2e5b252e165b5f40fae61d5">77fefdaffd4ef3cef2e5b252e165b5f40fae61d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.1.0","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;firefox-profile:0.2.7;archiver:0.7.1;file-utils:0.1.5;lodash:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.2.0","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;istanbul-middleware:0.2.2;archiver:0.14.4;lodash:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;wd:0.3.11;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"zuul:3.12.0;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file superagent package json path to vulnerable library superagent node modules file utils node modules lodash package json dependency hierarchy zuul tgz root library firefox profile tgz archiver tgz file utils tgz x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file superagent package json path to vulnerable library superagent node modules istanbul middleware node modules lodash package json dependency hierarchy zuul tgz root library istanbul middleware tgz archiver tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file superagent package json path to vulnerable library superagent node modules archiver node modules lodash package json superagent node modules findup sync node modules lodash package json superagent node modules wd node modules lodash package json dependency hierarchy zuul tgz root library wd tgz x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file superagent package json path to vulnerable library superagent node modules zuul node modules lodash package json dependency hierarchy zuul tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is vulnerabilityurl
0
79,350
28,120,002,104
IssuesEvent
2023-03-31 13:36:30
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: RemoteWebDriver failed while trying to upload file
I-defect needs-triaging
### What happened? I am trying to upload the file to the web application. The file size is like 4 gigabytes (that's the point, I guess). My script looks like this: ``` python elem = self.driver.find_element("//input[@type='file']") elem.send_keys(f"/tmp/filename") elem = self.driver.find_element("//button[@data-testid='upload-btn']") elem.click() ``` ### How can we reproduce the issue? ```shell Upload large file via RemoteWebDriver. ``` ### Relevant log output ```shell Traceback (most recent call last): File "/app/file.py", line 115, in <module> scenario.run() File "/app/file.py", line 74, in run elem.send_keys(f"/tmp/{distributable_name}") File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 475, in send_keys value = self._upload(local_file) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 708, in _upload raise e File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 699, in _upload return self._execute(Command.UPLOAD_FILE, {'file': content})['value'] File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 633, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webdriver.py", line 321, in execute self.error_handler.check_response(response) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/errorhandler.py", line 242, in check_response raise exception_class(message, screen, stacktrace) selenium.common.exceptions.WebDriverException: Message: None Stacktrace: at java.io.ByteArrayOutputStream.hugeCapacity (ByteArrayOutputStream.java:123) at java.io.ByteArrayOutputStream.grow (ByteArrayOutputStream.java:117) at java.io.ByteArrayOutputStream.ensureCapacity (ByteArrayOutputStream.java:93) at java.io.ByteArrayOutputStream.write (ByteArrayOutputStream.java:153) at com.google.common.io.ByteStreams.copy (ByteStreams.java:113) at org.openqa.selenium.remote.http.HttpMessage.getContent (HttpMessage.java:140) at org.openqa.selenium.remote.http.HttpRequest.getContent (HttpRequest.java:25) at org.openqa.selenium.remote.http.HttpMessage.getContentString (HttpMessage.java:152) at org.openqa.selenium.remote.http.HttpRequest.getContentString (HttpRequest.java:25) at org.openqa.selenium.remote.server.commandhandler.UploadFile.execute (UploadFile.java:47) at org.openqa.selenium.remote.server.WebDriverServlet.lambda$handle$0 (WebDriverServlet.java:235) at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:511) at java.util.concurrent.FutureTask.run (FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:624) at java.lang.Thread.run (Thread.java:748) ``` ### Operating System Ubuntu ### Selenium version 4.141.0 ### What are the browser(s) and version(s) where you see this issue? selenium/standalone-chrome-debug ### What are the browser driver(s) and version(s) where you see this issue? selenium/standalone-chrome-debug ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: RemoteWebDriver failed while trying to upload file - ### What happened? I am trying to upload the file to the web application. The file size is like 4 gigabytes (that's the point, I guess). My script looks like this: ``` python elem = self.driver.find_element("//input[@type='file']") elem.send_keys(f"/tmp/filename") elem = self.driver.find_element("//button[@data-testid='upload-btn']") elem.click() ``` ### How can we reproduce the issue? ```shell Upload large file via RemoteWebDriver. ``` ### Relevant log output ```shell Traceback (most recent call last): File "/app/file.py", line 115, in <module> scenario.run() File "/app/file.py", line 74, in run elem.send_keys(f"/tmp/{distributable_name}") File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 475, in send_keys value = self._upload(local_file) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 708, in _upload raise e File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 699, in _upload return self._execute(Command.UPLOAD_FILE, {'file': content})['value'] File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webelement.py", line 633, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/webdriver.py", line 321, in execute self.error_handler.check_response(response) File "/usr/local/lib/python3.8/dist-packages/selenium/webdriver/remote/errorhandler.py", line 242, in check_response raise exception_class(message, screen, stacktrace) selenium.common.exceptions.WebDriverException: Message: None Stacktrace: at java.io.ByteArrayOutputStream.hugeCapacity (ByteArrayOutputStream.java:123) at java.io.ByteArrayOutputStream.grow (ByteArrayOutputStream.java:117) at java.io.ByteArrayOutputStream.ensureCapacity (ByteArrayOutputStream.java:93) at java.io.ByteArrayOutputStream.write (ByteArrayOutputStream.java:153) at com.google.common.io.ByteStreams.copy (ByteStreams.java:113) at org.openqa.selenium.remote.http.HttpMessage.getContent (HttpMessage.java:140) at org.openqa.selenium.remote.http.HttpRequest.getContent (HttpRequest.java:25) at org.openqa.selenium.remote.http.HttpMessage.getContentString (HttpMessage.java:152) at org.openqa.selenium.remote.http.HttpRequest.getContentString (HttpRequest.java:25) at org.openqa.selenium.remote.server.commandhandler.UploadFile.execute (UploadFile.java:47) at org.openqa.selenium.remote.server.WebDriverServlet.lambda$handle$0 (WebDriverServlet.java:235) at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:511) at java.util.concurrent.FutureTask.run (FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:624) at java.lang.Thread.run (Thread.java:748) ``` ### Operating System Ubuntu ### Selenium version 4.141.0 ### What are the browser(s) and version(s) where you see this issue? selenium/standalone-chrome-debug ### What are the browser driver(s) and version(s) where you see this issue? selenium/standalone-chrome-debug ### Are you using Selenium Grid? _No response_
non_infrastructure
remotewebdriver failed while trying to upload file what happened i am trying to upload the file to the web application the file size is like gigabytes that s the point i guess my script looks like this python elem self driver find element input elem send keys f tmp filename elem self driver find element button elem click how can we reproduce the issue shell upload large file via remotewebdriver relevant log output shell traceback most recent call last file app file py line in scenario run file app file py line in run elem send keys f tmp distributable name file usr local lib dist packages selenium webdriver remote webelement py line in send keys value self upload local file file usr local lib dist packages selenium webdriver remote webelement py line in upload raise e file usr local lib dist packages selenium webdriver remote webelement py line in upload return self execute command upload file file content file usr local lib dist packages selenium webdriver remote webelement py line in execute return self parent execute command params file usr local lib dist packages selenium webdriver remote webdriver py line in execute self error handler check response response file usr local lib dist packages selenium webdriver remote errorhandler py line in check response raise exception class message screen stacktrace selenium common exceptions webdriverexception message none stacktrace at java io bytearrayoutputstream hugecapacity bytearrayoutputstream java at java io bytearrayoutputstream grow bytearrayoutputstream java at java io bytearrayoutputstream ensurecapacity bytearrayoutputstream java at java io bytearrayoutputstream write bytearrayoutputstream java at com google common io bytestreams copy bytestreams java at org openqa selenium remote http httpmessage getcontent httpmessage java at org openqa selenium remote http httprequest getcontent httprequest java at org openqa selenium remote http httpmessage getcontentstring httpmessage java at org openqa selenium remote http httprequest getcontentstring httprequest java at org openqa selenium remote server commandhandler uploadfile execute uploadfile java at org openqa selenium remote server webdriverservlet lambda handle webdriverservlet java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java operating system ubuntu selenium version what are the browser s and version s where you see this issue selenium standalone chrome debug what are the browser driver s and version s where you see this issue selenium standalone chrome debug are you using selenium grid no response
0
92,354
8,360,986,032
IssuesEvent
2018-10-03 13:14:19
mF2C/mF2C
https://api.github.com/repos/mF2C/mF2C
closed
A way to provide service/COMPSs job results and status to calling clients
TESTING AND UCs demo
An mF2C service's developer can create COMPSs jobs, then register and run them via the lifecycle component. This is also true for docker/dcompose services, but let's just focus on COMPSs jobs. Once the job is submitted---and, to prevent misunderstandings, this means executing e.g. Java code that a developer has written in the COMPSs framework---one currently has no way of interfacing with the job. Developers are somewhat expected to provide or create their own result distribution framework. This is something mF2C as a framework should provide. During the runtime of the service, and after it, there should exist an API call to either lifecycle, cimi or another service, whichever is deemed the proper way to interface with mF2C, with which a developer can qeury the status of the submitted job. Even basic statuses would be fine, e.g. created, running and finished. Once a job is done, results need to be extracted from it in order to use them. In UC2, we currently access and insert results into the (local) application database directly, which is not desired in distributed systems. As we have submitted the job via a lifecycle endpoint, one would expect to have results available, on completion, at another endpoint of the same or related service. The challenges for this are that, theoretically, the lifecycle component can start the job in a completely unrelated node where no components of an _mF2C application_ are even running. The system would thus need to support the efficient transfer of results to the request-initiating node. All details should be abstracted away from the developer via API calls to a service. Another challenge is the storage of results - how long are they stored for? With what priority compared to other results? In the not-unlikely event that the node containing the result suffers from sudden destruction or perhaps just a loss of connectivity, will the computation data be stored redundantly? What would this redundancy entail with regards to being resilient to different magnitudes of network graph cuts? Would the computation be repeated? Would all this be configurable?
1.0
A way to provide service/COMPSs job results and status to calling clients - An mF2C service's developer can create COMPSs jobs, then register and run them via the lifecycle component. This is also true for docker/dcompose services, but let's just focus on COMPSs jobs. Once the job is submitted---and, to prevent misunderstandings, this means executing e.g. Java code that a developer has written in the COMPSs framework---one currently has no way of interfacing with the job. Developers are somewhat expected to provide or create their own result distribution framework. This is something mF2C as a framework should provide. During the runtime of the service, and after it, there should exist an API call to either lifecycle, cimi or another service, whichever is deemed the proper way to interface with mF2C, with which a developer can qeury the status of the submitted job. Even basic statuses would be fine, e.g. created, running and finished. Once a job is done, results need to be extracted from it in order to use them. In UC2, we currently access and insert results into the (local) application database directly, which is not desired in distributed systems. As we have submitted the job via a lifecycle endpoint, one would expect to have results available, on completion, at another endpoint of the same or related service. The challenges for this are that, theoretically, the lifecycle component can start the job in a completely unrelated node where no components of an _mF2C application_ are even running. The system would thus need to support the efficient transfer of results to the request-initiating node. All details should be abstracted away from the developer via API calls to a service. Another challenge is the storage of results - how long are they stored for? With what priority compared to other results? In the not-unlikely event that the node containing the result suffers from sudden destruction or perhaps just a loss of connectivity, will the computation data be stored redundantly? What would this redundancy entail with regards to being resilient to different magnitudes of network graph cuts? Would the computation be repeated? Would all this be configurable?
non_infrastructure
a way to provide service compss job results and status to calling clients an service s developer can create compss jobs then register and run them via the lifecycle component this is also true for docker dcompose services but let s just focus on compss jobs once the job is submitted and to prevent misunderstandings this means executing e g java code that a developer has written in the compss framework one currently has no way of interfacing with the job developers are somewhat expected to provide or create their own result distribution framework this is something as a framework should provide during the runtime of the service and after it there should exist an api call to either lifecycle cimi or another service whichever is deemed the proper way to interface with with which a developer can qeury the status of the submitted job even basic statuses would be fine e g created running and finished once a job is done results need to be extracted from it in order to use them in we currently access and insert results into the local application database directly which is not desired in distributed systems as we have submitted the job via a lifecycle endpoint one would expect to have results available on completion at another endpoint of the same or related service the challenges for this are that theoretically the lifecycle component can start the job in a completely unrelated node where no components of an application are even running the system would thus need to support the efficient transfer of results to the request initiating node all details should be abstracted away from the developer via api calls to a service another challenge is the storage of results how long are they stored for with what priority compared to other results in the not unlikely event that the node containing the result suffers from sudden destruction or perhaps just a loss of connectivity will the computation data be stored redundantly what would this redundancy entail with regards to being resilient to different magnitudes of network graph cuts would the computation be repeated would all this be configurable
0
16,578
12,058,029,460
IssuesEvent
2020-04-15 16:46:07
skypyproject/skypy
https://api.github.com/repos/skypyproject/skypy
closed
code file cleanup
good first issue help wanted infrastructure low priority v0.1 Hack
There are a couple leftovers from package creation, such as `skypy/example_mod.py` and `skypy/example_subpkg`. They have served their purpose and need to go for v0.1.
1.0
code file cleanup - There are a couple leftovers from package creation, such as `skypy/example_mod.py` and `skypy/example_subpkg`. They have served their purpose and need to go for v0.1.
infrastructure
code file cleanup there are a couple leftovers from package creation such as skypy example mod py and skypy example subpkg they have served their purpose and need to go for
1
79,501
28,314,214,023
IssuesEvent
2023-04-10 18:06:05
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
FE: Sidebar menus should not include links to unpublished nodes.
Defect VA.gov frontend Menus Facilities
## Describe the defect As a content editor I want the sidebar menus in my section of the site (primarily vamcs) to only include links to published items so that site visitors do not encounter broken liinks. The current solution if for a content admin to go into the menu and disable the menu item that links to the unpublished page. Then later when the page is published, they must remember to go in and enable the menu item. The module menu_graphQL was [added back in Sept 2021](https://github.com/department-of-veterans-affairs/va.gov-cms/pull/6568) to make it possible to get the published state of a node on menu item like this ![image](https://user-images.githubusercontent.com/5752113/194953143-be6f3aa7-a539-479e-aca7-d4439b5a1859.png) That information might not have been added yet to the menu queries but should be, and then used to prevent rendering menu links to items that are unpublished. ## AC / Expected behavior - [ ] When a node is unpublished (either draft or archived) the FE menu does not render a link to that page. - [ ] When a node is published and has an associated menu item, the menu item should be rendered with a link to the page. ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [x] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
FE: Sidebar menus should not include links to unpublished nodes. - ## Describe the defect As a content editor I want the sidebar menus in my section of the site (primarily vamcs) to only include links to published items so that site visitors do not encounter broken liinks. The current solution if for a content admin to go into the menu and disable the menu item that links to the unpublished page. Then later when the page is published, they must remember to go in and enable the menu item. The module menu_graphQL was [added back in Sept 2021](https://github.com/department-of-veterans-affairs/va.gov-cms/pull/6568) to make it possible to get the published state of a node on menu item like this ![image](https://user-images.githubusercontent.com/5752113/194953143-be6f3aa7-a539-479e-aca7-d4439b5a1859.png) That information might not have been added yet to the menu queries but should be, and then used to prevent rendering menu links to items that are unpublished. ## AC / Expected behavior - [ ] When a node is unpublished (either draft or archived) the FE menu does not render a link to that page. - [ ] When a node is published and has an associated menu item, the menu item should be rendered with a link to the page. ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [x] `⭐️ Facilities` - [ ] `⭐️ User support`
non_infrastructure
fe sidebar menus should not include links to unpublished nodes describe the defect as a content editor i want the sidebar menus in my section of the site primarily vamcs to only include links to published items so that site visitors do not encounter broken liinks the current solution if for a content admin to go into the menu and disable the menu item that links to the unpublished page then later when the page is published they must remember to go in and enable the menu item the module menu graphql was to make it possible to get the published state of a node on menu item like this that information might not have been added yet to the menu queries but should be and then used to prevent rendering menu links to items that are unpublished ac expected behavior when a node is unpublished either draft or archived the fe menu does not render a link to that page when a node is published and has an associated menu item the menu item should be rendered with a link to the page labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
0
82,032
10,267,548,546
IssuesEvent
2019-08-23 02:08:56
tamb/domponent
https://api.github.com/repos/tamb/domponent
closed
Feedback
documentation enhancement
I am working on a similar thing called [dom99](https://github.com/GrosSacASac/DOM99/) The docs are a bit out of date, but check the examples. After skimming through your project: data-state You basically evented yet another data format, consider instead using something native of the platform like data-state="{count:24, isEven:true}" or data-state-count="24" data-state-isEven="true" data-key it is optional, must be unique , but what does it do ? Component Fields why prefix things with $ ? do not commit domponent-1.0.0-alpha.18.tgz Otherwise very interesting project, I give more detailed feedback later
1.0
Feedback - I am working on a similar thing called [dom99](https://github.com/GrosSacASac/DOM99/) The docs are a bit out of date, but check the examples. After skimming through your project: data-state You basically evented yet another data format, consider instead using something native of the platform like data-state="{count:24, isEven:true}" or data-state-count="24" data-state-isEven="true" data-key it is optional, must be unique , but what does it do ? Component Fields why prefix things with $ ? do not commit domponent-1.0.0-alpha.18.tgz Otherwise very interesting project, I give more detailed feedback later
non_infrastructure
feedback i am working on a similar thing called the docs are a bit out of date but check the examples after skimming through your project data state you basically evented yet another data format consider instead using something native of the platform like data state count iseven true or data state count data state iseven true data key it is optional must be unique but what does it do component fields why prefix things with do not commit domponent alpha tgz otherwise very interesting project i give more detailed feedback later
0
87,846
8,124,054,136
IssuesEvent
2018-08-16 16:16:42
freedomofpress/securedrop
https://api.github.com/repos/freedomofpress/securedrop
opened
Add VNC server to functional test runner container
functional_tests
## Description It is very difficult currently to debug why functional tests fail using the new tor browser selenium logic. The upstream selenium team offers docker containers packed with firefox/chrome that include a VNC server for visual view of what the browser is doing. This would be insanely useful for us during development. ## User Research Evidence Doesn't affect end-users, only developers. ## User Stories As a developer, it would aid in my troubleshooting to be able to visually see the Tor Browser tests running while they are running.
1.0
Add VNC server to functional test runner container - ## Description It is very difficult currently to debug why functional tests fail using the new tor browser selenium logic. The upstream selenium team offers docker containers packed with firefox/chrome that include a VNC server for visual view of what the browser is doing. This would be insanely useful for us during development. ## User Research Evidence Doesn't affect end-users, only developers. ## User Stories As a developer, it would aid in my troubleshooting to be able to visually see the Tor Browser tests running while they are running.
non_infrastructure
add vnc server to functional test runner container description it is very difficult currently to debug why functional tests fail using the new tor browser selenium logic the upstream selenium team offers docker containers packed with firefox chrome that include a vnc server for visual view of what the browser is doing this would be insanely useful for us during development user research evidence doesn t affect end users only developers user stories as a developer it would aid in my troubleshooting to be able to visually see the tor browser tests running while they are running
0
621
2,794,248,921
IssuesEvent
2015-05-11 15:42:08
cooperative-computing-lab/cctools
https://api.github.com/repos/cooperative-computing-lab/cctools
closed
Autobuild Docker Images
bug infrastructure
Take a look at the docker images here: https://registry.hub.docker.com/repos/cclnd/ Note that `ubuntu` is a base ubuntu image while `cctools-env-full-ubuntu` is derived from ubuntu, with all of the necessary cctools dependencies installed. Same for `centos` and `debian` Now, builds on those platforms are currently failing tests because the platforms do not have some essential tools installed: `curl ImageMagick python perl`. So, we need to re-build those images with the missing packages. @vhawley Please post (right here would be fine) the Dockerfiles that you used to generate these images. We need to adjust them in order to generate new images...
1.0
Autobuild Docker Images - Take a look at the docker images here: https://registry.hub.docker.com/repos/cclnd/ Note that `ubuntu` is a base ubuntu image while `cctools-env-full-ubuntu` is derived from ubuntu, with all of the necessary cctools dependencies installed. Same for `centos` and `debian` Now, builds on those platforms are currently failing tests because the platforms do not have some essential tools installed: `curl ImageMagick python perl`. So, we need to re-build those images with the missing packages. @vhawley Please post (right here would be fine) the Dockerfiles that you used to generate these images. We need to adjust them in order to generate new images...
infrastructure
autobuild docker images take a look at the docker images here note that ubuntu is a base ubuntu image while cctools env full ubuntu is derived from ubuntu with all of the necessary cctools dependencies installed same for centos and debian now builds on those platforms are currently failing tests because the platforms do not have some essential tools installed curl imagemagick python perl so we need to re build those images with the missing packages vhawley please post right here would be fine the dockerfiles that you used to generate these images we need to adjust them in order to generate new images
1
206,535
23,390,840,609
IssuesEvent
2022-08-11 17:40:26
jgeraigery/HubSpot-Test-Repo-2
https://api.github.com/repos/jgeraigery/HubSpot-Test-Repo-2
opened
dropwizard-configuration-1.3.12.jar: 2 vulnerabilities (highest severity is: 8.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-configuration-1.3.12.jar</b></p></summary> <p></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11002](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11002) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | dropwizard-validation-1.3.12.jar | Transitive | 1.3.21 | &#9989; | | [CVE-2020-5245](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5245) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | dropwizard-validation-1.3.12.jar | Transitive | 1.3.19 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-11002</summary> ### Vulnerable Library - <b>dropwizard-validation-1.3.12.jar</b></p> <p>Dropwizard is a Java framework for developing ops-friendly, high-performance, RESTful web applications.</p> <p>Library home page: <a href="http://www.dropwizard.io/1.3.12">http://www.dropwizard.io/1.3.12</a></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> Dependency Hierarchy: - dropwizard-configuration-1.3.12.jar (Root Library) - :x: **dropwizard-validation-1.3.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> dropwizard-validation before versions 2.0.3 and 1.3.21 has a remote code execution vulnerability. A server-side template injection was identified in the self-validating feature enabling attackers to inject arbitrary Java EL expressions, leading to Remote Code Execution (RCE) vulnerability. If you are using a self-validating bean an upgrade to Dropwizard 1.3.21/2.0.3 or later is strongly recommended. The changes introduced in Dropwizard 1.3.19 and 2.0.2 for CVE-2020-5245 unfortunately did not fix the underlying issue completely. The issue has been fixed in dropwizard-validation 1.3.21 and 2.0.3 or later. We strongly recommend upgrading to one of these versions. <p>Publish Date: 2020-04-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11002>CVE-2020-11002</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/dropwizard/dropwizard/security/advisories/GHSA-8jpx-m2wh-2v34">https://github.com/dropwizard/dropwizard/security/advisories/GHSA-8jpx-m2wh-2v34</a></p> <p>Release Date: 2020-04-13</p> <p>Fix Resolution (io.dropwizard:dropwizard-validation): 1.3.21</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-configuration): 1.3.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-5245</summary> ### Vulnerable Library - <b>dropwizard-validation-1.3.12.jar</b></p> <p>Dropwizard is a Java framework for developing ops-friendly, high-performance, RESTful web applications.</p> <p>Library home page: <a href="http://www.dropwizard.io/1.3.12">http://www.dropwizard.io/1.3.12</a></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> Dependency Hierarchy: - dropwizard-configuration-1.3.12.jar (Root Library) - :x: **dropwizard-validation-1.3.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Dropwizard-Validation before 1.3.19, and 2.0.2 may allow arbitrary code execution on the host system, with the privileges of the Dropwizard service account, by injecting arbitrary Java Expression Language expressions when using the self-validating feature. The issue has been fixed in dropwizard-validation 1.3.19 and 2.0.2. <p>Publish Date: 2020-02-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5245>CVE-2020-5245</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-5245">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-5245</a></p> <p>Release Date: 2020-02-24</p> <p>Fix Resolution (io.dropwizard:dropwizard-validation): 1.3.19</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-configuration): 1.3.19</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
dropwizard-configuration-1.3.12.jar: 2 vulnerabilities (highest severity is: 8.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-configuration-1.3.12.jar</b></p></summary> <p></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11002](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11002) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | dropwizard-validation-1.3.12.jar | Transitive | 1.3.21 | &#9989; | | [CVE-2020-5245](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5245) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | dropwizard-validation-1.3.12.jar | Transitive | 1.3.19 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-11002</summary> ### Vulnerable Library - <b>dropwizard-validation-1.3.12.jar</b></p> <p>Dropwizard is a Java framework for developing ops-friendly, high-performance, RESTful web applications.</p> <p>Library home page: <a href="http://www.dropwizard.io/1.3.12">http://www.dropwizard.io/1.3.12</a></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> Dependency Hierarchy: - dropwizard-configuration-1.3.12.jar (Root Library) - :x: **dropwizard-validation-1.3.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> dropwizard-validation before versions 2.0.3 and 1.3.21 has a remote code execution vulnerability. A server-side template injection was identified in the self-validating feature enabling attackers to inject arbitrary Java EL expressions, leading to Remote Code Execution (RCE) vulnerability. If you are using a self-validating bean an upgrade to Dropwizard 1.3.21/2.0.3 or later is strongly recommended. The changes introduced in Dropwizard 1.3.19 and 2.0.2 for CVE-2020-5245 unfortunately did not fix the underlying issue completely. The issue has been fixed in dropwizard-validation 1.3.21 and 2.0.3 or later. We strongly recommend upgrading to one of these versions. <p>Publish Date: 2020-04-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11002>CVE-2020-11002</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/dropwizard/dropwizard/security/advisories/GHSA-8jpx-m2wh-2v34">https://github.com/dropwizard/dropwizard/security/advisories/GHSA-8jpx-m2wh-2v34</a></p> <p>Release Date: 2020-04-13</p> <p>Fix Resolution (io.dropwizard:dropwizard-validation): 1.3.21</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-configuration): 1.3.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-5245</summary> ### Vulnerable Library - <b>dropwizard-validation-1.3.12.jar</b></p> <p>Dropwizard is a Java framework for developing ops-friendly, high-performance, RESTful web applications.</p> <p>Library home page: <a href="http://www.dropwizard.io/1.3.12">http://www.dropwizard.io/1.3.12</a></p> <p>Path to dependency file: /SingularityS3Uploader/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar,/home/wss-scanner/.m2/repository/io/dropwizard/dropwizard-validation/1.3.12/dropwizard-validation-1.3.12.jar</p> <p> Dependency Hierarchy: - dropwizard-configuration-1.3.12.jar (Root Library) - :x: **dropwizard-validation-1.3.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/HubSpot-Test-Repo-2/commit/c8f89b268c9e5524a7490639bbb63d2a9d40bd6c">c8f89b268c9e5524a7490639bbb63d2a9d40bd6c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Dropwizard-Validation before 1.3.19, and 2.0.2 may allow arbitrary code execution on the host system, with the privileges of the Dropwizard service account, by injecting arbitrary Java Expression Language expressions when using the self-validating feature. The issue has been fixed in dropwizard-validation 1.3.19 and 2.0.2. <p>Publish Date: 2020-02-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5245>CVE-2020-5245</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-5245">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-5245</a></p> <p>Release Date: 2020-02-24</p> <p>Fix Resolution (io.dropwizard:dropwizard-validation): 1.3.19</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-configuration): 1.3.19</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_infrastructure
dropwizard configuration jar vulnerabilities highest severity is vulnerable library dropwizard configuration jar path to dependency file pom xml path to vulnerable library home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high dropwizard validation jar transitive high dropwizard validation jar transitive details cve vulnerable library dropwizard validation jar dropwizard is a java framework for developing ops friendly high performance restful web applications library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar dependency hierarchy dropwizard configuration jar root library x dropwizard validation jar vulnerable library found in head commit a href found in base branch master vulnerability details dropwizard validation before versions and has a remote code execution vulnerability a server side template injection was identified in the self validating feature enabling attackers to inject arbitrary java el expressions leading to remote code execution rce vulnerability if you are using a self validating bean an upgrade to dropwizard or later is strongly recommended the changes introduced in dropwizard and for cve unfortunately did not fix the underlying issue completely the issue has been fixed in dropwizard validation and or later we strongly recommend upgrading to one of these versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io dropwizard dropwizard validation direct dependency fix resolution io dropwizard dropwizard configuration rescue worker helmet automatic remediation is available for this issue cve vulnerable library dropwizard validation jar dropwizard is a java framework for developing ops friendly high performance restful web applications library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar home wss scanner repository io dropwizard dropwizard validation dropwizard validation jar dependency hierarchy dropwizard configuration jar root library x dropwizard validation jar vulnerable library found in head commit a href found in base branch master vulnerability details dropwizard validation before and may allow arbitrary code execution on the host system with the privileges of the dropwizard service account by injecting arbitrary java expression language expressions when using the self validating feature the issue has been fixed in dropwizard validation and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io dropwizard dropwizard validation direct dependency fix resolution io dropwizard dropwizard configuration rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
65,208
19,272,276,256
IssuesEvent
2021-12-10 07:34:50
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
introduce default device for tiered-store configuration
Type: Defect
Introduce a (local) default device for tiered-store configuration, so that users can enable tiered-store without configuring a device
1.0
introduce default device for tiered-store configuration - Introduce a (local) default device for tiered-store configuration, so that users can enable tiered-store without configuring a device
non_infrastructure
introduce default device for tiered store configuration introduce a local default device for tiered store configuration so that users can enable tiered store without configuring a device
0
7,465
6,965,340,227
IssuesEvent
2017-12-09 04:47:48
nix-rust/nix
https://api.github.com/repos/nix-rust/nix
closed
TravisCI is too slow
A-infrastructure
We have 34 builds in Travis, and it takes over an hour for all of them to finish. It's causing Bors to timeout. At a minimum, we need to increase bors's timeout. Other ways to speed up the builds include: 1. Move i686-unknown-freebsd into BuildBot 2. remove `sudo: required`. It causes Travis to use a full VM instead of a container, and I don't think we need it. 3. Only test beta and nightly builds on one architecture. 4. Enable parallel tests. This will likely have minimal benefit, because setting up the environment is slower than running tests. 5. Use a custom Docker image so we won't have to download the QEMU stuff all the time 6. Set Travis's Auto Cancel Pull Requests feature Any other ideas?
1.0
TravisCI is too slow - We have 34 builds in Travis, and it takes over an hour for all of them to finish. It's causing Bors to timeout. At a minimum, we need to increase bors's timeout. Other ways to speed up the builds include: 1. Move i686-unknown-freebsd into BuildBot 2. remove `sudo: required`. It causes Travis to use a full VM instead of a container, and I don't think we need it. 3. Only test beta and nightly builds on one architecture. 4. Enable parallel tests. This will likely have minimal benefit, because setting up the environment is slower than running tests. 5. Use a custom Docker image so we won't have to download the QEMU stuff all the time 6. Set Travis's Auto Cancel Pull Requests feature Any other ideas?
infrastructure
travisci is too slow we have builds in travis and it takes over an hour for all of them to finish it s causing bors to timeout at a minimum we need to increase bors s timeout other ways to speed up the builds include move unknown freebsd into buildbot remove sudo required it causes travis to use a full vm instead of a container and i don t think we need it only test beta and nightly builds on one architecture enable parallel tests this will likely have minimal benefit because setting up the environment is slower than running tests use a custom docker image so we won t have to download the qemu stuff all the time set travis s auto cancel pull requests feature any other ideas
1
378,241
11,199,264,974
IssuesEvent
2020-01-03 18:13:13
containrrr/watchtower
https://api.github.com/repos/containrrr/watchtower
closed
Unable to update container when a container raise error
Priority: Medium Status: Available Type: Bug
sorry my english is not well **Describe the bug** i have 3 containers # docker-compose.yml ``` version: "3" services: aa: image: registry-xxxx.com/xxx/aa restart: always expose: - 8000/tcp bb: image: registry-xxxx.com/xxx/bb restart: always expose: - 8009/tcp cc: image: registry-xxxx.com/xxx/cc restart: always expose: - 8005/tcp watchtower: image: containrrr/watchtower volumes: - /var/run/docker.sock:/var/run/docker.sock - /home/aaa/.docker/config.json:/config.json command: --interval 30 ``` when one container raise error, then this container can't to update **To Reproduce** Steps to reproduce the behavior: 1. when i push a new bad code to container aa (the new bad code will be raise a error) **like this:** ![image](https://user-images.githubusercontent.com/6580897/68990803-6fabd100-0892-11ea-893e-98b0fe50d4ed.png) 2. then i push a new right code to container aa (the new code is right), but watchtower can't update the container aa **Expected behavior** i hope that when container raise error, then when i push a new right code , watchtower can to get new container **Environment** Client: Debug Mode: false Server: Containers: 5 Running: 0 Paused: 0 Stopped: 5 Images: 8 Server Version: 19.03.4 Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Native Overlay Diff: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: b34a5c8af56e510852c35414db4c1f4fa6172339 runc version: 3e425f80a8c931f88e6d94a8c831b9d5aa481657 init version: fec3683 Security Options: apparmor seccomp Profile: default Kernel Version: 4.15.0-54-generic Operating System: Ubuntu 18.04.1 LTS OSType: linux Architecture: x86_64 chinese version: 你好, 我有3个容器,分别是aa,bb,cc 一开始这3个容器工作的很好 但是我当推送了错误的代码到容器aa,导致容器aa 因为错误而退出了 接着,我立刻推送了新的正确代码到容器aa,可是 watchtower 却无法 拉取到新的容器aa 我希望,当容器aa遇到错误代码崩溃后,还是可以在后台不断请求 新的容器aa 因为在新的容器aa上,我解决了 错误
1.0
Unable to update container when a container raise error - sorry my english is not well **Describe the bug** i have 3 containers # docker-compose.yml ``` version: "3" services: aa: image: registry-xxxx.com/xxx/aa restart: always expose: - 8000/tcp bb: image: registry-xxxx.com/xxx/bb restart: always expose: - 8009/tcp cc: image: registry-xxxx.com/xxx/cc restart: always expose: - 8005/tcp watchtower: image: containrrr/watchtower volumes: - /var/run/docker.sock:/var/run/docker.sock - /home/aaa/.docker/config.json:/config.json command: --interval 30 ``` when one container raise error, then this container can't to update **To Reproduce** Steps to reproduce the behavior: 1. when i push a new bad code to container aa (the new bad code will be raise a error) **like this:** ![image](https://user-images.githubusercontent.com/6580897/68990803-6fabd100-0892-11ea-893e-98b0fe50d4ed.png) 2. then i push a new right code to container aa (the new code is right), but watchtower can't update the container aa **Expected behavior** i hope that when container raise error, then when i push a new right code , watchtower can to get new container **Environment** Client: Debug Mode: false Server: Containers: 5 Running: 0 Paused: 0 Stopped: 5 Images: 8 Server Version: 19.03.4 Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Native Overlay Diff: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: b34a5c8af56e510852c35414db4c1f4fa6172339 runc version: 3e425f80a8c931f88e6d94a8c831b9d5aa481657 init version: fec3683 Security Options: apparmor seccomp Profile: default Kernel Version: 4.15.0-54-generic Operating System: Ubuntu 18.04.1 LTS OSType: linux Architecture: x86_64 chinese version: 你好, 我有3个容器,分别是aa,bb,cc 一开始这3个容器工作的很好 但是我当推送了错误的代码到容器aa,导致容器aa 因为错误而退出了 接着,我立刻推送了新的正确代码到容器aa,可是 watchtower 却无法 拉取到新的容器aa 我希望,当容器aa遇到错误代码崩溃后,还是可以在后台不断请求 新的容器aa 因为在新的容器aa上,我解决了 错误
non_infrastructure
unable to update container when a container raise error sorry my english is not well describe the bug i have containers docker compose yml version services aa image registry xxxx com xxx aa restart always expose tcp bb image registry xxxx com xxx bb restart always expose tcp cc image registry xxxx com xxx cc restart always expose tcp watchtower image containrrr watchtower volumes var run docker sock var run docker sock home aaa docker config json config json command interval when one container raise error then this container can t to update to reproduce steps to reproduce the behavior when i push a new bad code to container aa the new bad code will be raise a error like this then i push a new right code to container aa the new code is right but watchtower can t update the container aa expected behavior i hope that when container raise error then when i push a new right code watchtower can to get new container environment client debug mode false server containers running paused stopped images server version storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host ipvlan macvlan null overlay log awslogs fluentd gcplogs gelf journald json file local logentries splunk syslog swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options apparmor seccomp profile default kernel version generic operating system ubuntu lts ostype linux architecture chinese version 你好, ,分别是aa,bb,cc 但是我当推送了错误的代码到容器aa,导致容器aa 因为错误而退出了 接着,我立刻推送了新的正确代码到容器aa,可是 watchtower 却无法 拉取到新的容器aa 我希望,当容器aa遇到错误代码崩溃后,还是可以在后台不断请求 新的容器aa 因为在新的容器aa上,我解决了 错误
0
29,801
24,283,960,539
IssuesEvent
2022-09-28 20:04:51
OpenLiberty/openliberty.io
https://api.github.com/repos/OpenLiberty/openliberty.io
closed
Send Slack message after health check in Code Engine toolchain
infrastructure
We need to move the Slack message Tekton Task to be after the application health check task.
1.0
Send Slack message after health check in Code Engine toolchain - We need to move the Slack message Tekton Task to be after the application health check task.
infrastructure
send slack message after health check in code engine toolchain we need to move the slack message tekton task to be after the application health check task
1
26,677
20,508,937,889
IssuesEvent
2022-03-01 02:54:32
google/oss-fuzz
https://api.github.com/repos/google/oss-fuzz
reopened
CoverageReportIntegrationTest::test_coverage_report is failing
infrastructure priority
From e.g. https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true ``` =================================== FAILURES =================================== ______________ CoverageReportIntegrationTest.test_coverage_report ______________ [gw0] linux -- Python 3.8.12 /opt/hostedtoolcache/Python/3.8.12/x64/bin/python self = <run_fuzzers_test.CoverageReportIntegrationTest testMethod=test_coverage_report> _ = <MagicMock name='_upload_artifact_with_upload_js' id='140482468299728'> @mock.patch('filestore.github_actions._upload_artifact_with_upload_js') def test_coverage_report(self, _): """Tests generation of coverage reports end-to-end, from building to generation.""" with test_helpers.docker_temp_dir() as temp_dir: shared = os.path.join(temp_dir, 'shared') os.mkdir(shared) copy_command = ('cp -r /opt/code_coverage /shared && ' 'cp $(which llvm-profdata) /shared && ' 'cp $(which llvm-cov) /shared') assert helper.docker_run([ '-v', f'{shared}:/shared', 'gcr.io/oss-fuzz-base/base-runner', 'bash', '-c', copy_command ]) os.environ['CODE_COVERAGE_SRC'] = os.path.join(shared, 'code_coverage') os.environ['PATH'] += os.pathsep + shared # Do coverage build. build_config = test_helpers.create_build_config( oss_fuzz_project_name=EXAMPLE_PROJECT, project_repo_name='oss-fuzz', workspace=temp_dir, git_sha='0b95fe1039ed7c38fea1f97078316bfc1030c[523](https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true#step:9:523)', base_commit='da07464[524](https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true#step:9:524)33dc18bae699e355a9821285d863c8', sanitizer=self.SANITIZER, cfl_platform='github', # Needed for test not to fail because of permissions issues. bad_build_check=False) self.assertTrue(build_fuzzers.build_fuzzers(build_config)) # TODO(metzman): Get rid of this here and make 'compile' do this. chmod_command = ('chmod -R +r /out && ' 'find /out -type d -exec chmod +x {} +') assert helper.docker_run([ '-v', f'{os.path.join(temp_dir, "build-out")}:/out', 'gcr.io/oss-fuzz-base/base-builder', 'bash', '-c', chmod_command ]) # Generate report. run_config = test_helpers.create_run_config(fuzz_seconds=FUZZ_SECONDS, workspace=temp_dir, sanitizer=self.SANITIZER, mode='coverage', cfl_platform='github') result = run_fuzzers.run_fuzzers(run_config) self.assertEqual(result, run_fuzzers.RunFuzzersResult.NO_BUG_FOUND) expected_summary_path = os.path.join( TEST_DATA_PATH, 'example_coverage_report_summary.json') with open(expected_summary_path) as file_handle: expected_summary = json.loads(file_handle.read()) actual_summary_path = os.path.join(temp_dir, 'cifuzz-coverage', 'report', 'linux', 'summary.json') > with open(actual_summary_path) as file_handle: E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpfqnfmle8/cifuzz-coverage/report/linux/summary.json' ``` Related error: ``` ------------------------------ Captured log call ------------------------------- ERROR root:github_api.py:75 Request to https://api.github.com/repos/None/None/actions/artifacts?per_page=100&page=1 failed. Code: 401. Response: {'message': 'Bad credentials', 'documentation_url': 'https://docs.github.com/rest'} ERROR root:clusterfuzz_deployment.py:143 Failed to download corpus for target: do_stuff_fuzzer. Error: Github API request failed. =========================== short test summary info ============================ FAILED infra/cifuzz/run_fuzzers_test.py::CoverageReportIntegrationTest::test_coverage_report```
1.0
CoverageReportIntegrationTest::test_coverage_report is failing - From e.g. https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true ``` =================================== FAILURES =================================== ______________ CoverageReportIntegrationTest.test_coverage_report ______________ [gw0] linux -- Python 3.8.12 /opt/hostedtoolcache/Python/3.8.12/x64/bin/python self = <run_fuzzers_test.CoverageReportIntegrationTest testMethod=test_coverage_report> _ = <MagicMock name='_upload_artifact_with_upload_js' id='140482468299728'> @mock.patch('filestore.github_actions._upload_artifact_with_upload_js') def test_coverage_report(self, _): """Tests generation of coverage reports end-to-end, from building to generation.""" with test_helpers.docker_temp_dir() as temp_dir: shared = os.path.join(temp_dir, 'shared') os.mkdir(shared) copy_command = ('cp -r /opt/code_coverage /shared && ' 'cp $(which llvm-profdata) /shared && ' 'cp $(which llvm-cov) /shared') assert helper.docker_run([ '-v', f'{shared}:/shared', 'gcr.io/oss-fuzz-base/base-runner', 'bash', '-c', copy_command ]) os.environ['CODE_COVERAGE_SRC'] = os.path.join(shared, 'code_coverage') os.environ['PATH'] += os.pathsep + shared # Do coverage build. build_config = test_helpers.create_build_config( oss_fuzz_project_name=EXAMPLE_PROJECT, project_repo_name='oss-fuzz', workspace=temp_dir, git_sha='0b95fe1039ed7c38fea1f97078316bfc1030c[523](https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true#step:9:523)', base_commit='da07464[524](https://github.com/google/oss-fuzz/runs/5309713893?check_suite_focus=true#step:9:524)33dc18bae699e355a9821285d863c8', sanitizer=self.SANITIZER, cfl_platform='github', # Needed for test not to fail because of permissions issues. bad_build_check=False) self.assertTrue(build_fuzzers.build_fuzzers(build_config)) # TODO(metzman): Get rid of this here and make 'compile' do this. chmod_command = ('chmod -R +r /out && ' 'find /out -type d -exec chmod +x {} +') assert helper.docker_run([ '-v', f'{os.path.join(temp_dir, "build-out")}:/out', 'gcr.io/oss-fuzz-base/base-builder', 'bash', '-c', chmod_command ]) # Generate report. run_config = test_helpers.create_run_config(fuzz_seconds=FUZZ_SECONDS, workspace=temp_dir, sanitizer=self.SANITIZER, mode='coverage', cfl_platform='github') result = run_fuzzers.run_fuzzers(run_config) self.assertEqual(result, run_fuzzers.RunFuzzersResult.NO_BUG_FOUND) expected_summary_path = os.path.join( TEST_DATA_PATH, 'example_coverage_report_summary.json') with open(expected_summary_path) as file_handle: expected_summary = json.loads(file_handle.read()) actual_summary_path = os.path.join(temp_dir, 'cifuzz-coverage', 'report', 'linux', 'summary.json') > with open(actual_summary_path) as file_handle: E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpfqnfmle8/cifuzz-coverage/report/linux/summary.json' ``` Related error: ``` ------------------------------ Captured log call ------------------------------- ERROR root:github_api.py:75 Request to https://api.github.com/repos/None/None/actions/artifacts?per_page=100&page=1 failed. Code: 401. Response: {'message': 'Bad credentials', 'documentation_url': 'https://docs.github.com/rest'} ERROR root:clusterfuzz_deployment.py:143 Failed to download corpus for target: do_stuff_fuzzer. Error: Github API request failed. =========================== short test summary info ============================ FAILED infra/cifuzz/run_fuzzers_test.py::CoverageReportIntegrationTest::test_coverage_report```
infrastructure
coveragereportintegrationtest test coverage report is failing from e g failures coveragereportintegrationtest test coverage report linux python opt hostedtoolcache python bin python self mock patch filestore github actions upload artifact with upload js def test coverage report self tests generation of coverage reports end to end from building to generation with test helpers docker temp dir as temp dir shared os path join temp dir shared os mkdir shared copy command cp r opt code coverage shared cp which llvm profdata shared cp which llvm cov shared assert helper docker run v f shared shared gcr io oss fuzz base base runner bash c copy command os environ os path join shared code coverage os environ os pathsep shared do coverage build build config test helpers create build config oss fuzz project name example project project repo name oss fuzz workspace temp dir git sha base commit sanitizer self sanitizer cfl platform github needed for test not to fail because of permissions issues bad build check false self asserttrue build fuzzers build fuzzers build config todo metzman get rid of this here and make compile do this chmod command chmod r r out find out type d exec chmod x assert helper docker run v f os path join temp dir build out out gcr io oss fuzz base base builder bash c chmod command generate report run config test helpers create run config fuzz seconds fuzz seconds workspace temp dir sanitizer self sanitizer mode coverage cfl platform github result run fuzzers run fuzzers run config self assertequal result run fuzzers runfuzzersresult no bug found expected summary path os path join test data path example coverage report summary json with open expected summary path as file handle expected summary json loads file handle read actual summary path os path join temp dir cifuzz coverage report linux summary json with open actual summary path as file handle e filenotfounderror no such file or directory tmp cifuzz coverage report linux summary json related error captured log call error root github api py request to failed code response message bad credentials documentation url error root clusterfuzz deployment py failed to download corpus for target do stuff fuzzer error github api request failed short test summary info failed infra cifuzz run fuzzers test py coveragereportintegrationtest test coverage report
1
2,747
3,837,769,913
IssuesEvent
2016-04-02 00:28:59
nnnick/Chart.js
https://api.github.com/repos/nnnick/Chart.js
closed
Bug fix missing from 1.0.2 Release
Category: Infrastructure Priority: p0 Version: 1.x
In the release notes for 1.0.2 this is mentioned. > - Handle negative arc sizing in chart segments But the fix for this is missing in the actual release. Which was committed here https://github.com/nnnick/Chart.js/commit/1df2b37bc5b3f66ed38d53d4a12b96f1d3a4fb9a Could we possibly get a 1.0.3 release with this fix included?
1.0
Bug fix missing from 1.0.2 Release - In the release notes for 1.0.2 this is mentioned. > - Handle negative arc sizing in chart segments But the fix for this is missing in the actual release. Which was committed here https://github.com/nnnick/Chart.js/commit/1df2b37bc5b3f66ed38d53d4a12b96f1d3a4fb9a Could we possibly get a 1.0.3 release with this fix included?
infrastructure
bug fix missing from release in the release notes for this is mentioned handle negative arc sizing in chart segments but the fix for this is missing in the actual release which was committed here could we possibly get a release with this fix included
1
84,653
15,724,744,186
IssuesEvent
2021-03-29 09:10:07
crouchr/learnage
https://api.github.com/repos/crouchr/learnage
opened
CVE-2019-1010279 (High) detected in suricatasuricata-1.0.2
security vulnerability
## CVE-2019-1010279 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>suricatasuricata-1.0.2</b></p></summary> <p> <p>Suricata git repository maintained by the OISF</p> <p>Library home page: <a href=https://github.com/inliniac/suricata.git>https://github.com/inliniac/suricata.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/crouchr/learnage/commit/a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c">a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>learnage/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/detect.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Open Information Security Foundation Suricata prior to version 4.1.3 is affected by: Denial of Service - TCP/HTTP detection bypass. The impact is: An attacker can evade a signature detection with a specialy formed sequence of network packets. The component is: detect.c (https://github.com/OISF/suricata/pull/3625/commits/d8634daf74c882356659addb65fb142b738a186b). The attack vector is: An attacker can trigger the vulnerability by a specifically crafted network TCP session. The fixed version is: 4.1.3. <p>Publish Date: 2019-07-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010279>CVE-2019-1010279</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010279">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010279</a></p> <p>Release Date: 2019-07-18</p> <p>Fix Resolution: 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-1010279 (High) detected in suricatasuricata-1.0.2 - ## CVE-2019-1010279 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>suricatasuricata-1.0.2</b></p></summary> <p> <p>Suricata git repository maintained by the OISF</p> <p>Library home page: <a href=https://github.com/inliniac/suricata.git>https://github.com/inliniac/suricata.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/crouchr/learnage/commit/a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c">a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>learnage/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/detect.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Open Information Security Foundation Suricata prior to version 4.1.3 is affected by: Denial of Service - TCP/HTTP detection bypass. The impact is: An attacker can evade a signature detection with a specialy formed sequence of network packets. The component is: detect.c (https://github.com/OISF/suricata/pull/3625/commits/d8634daf74c882356659addb65fb142b738a186b). The attack vector is: An attacker can trigger the vulnerability by a specifically crafted network TCP session. The fixed version is: 4.1.3. <p>Publish Date: 2019-07-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010279>CVE-2019-1010279</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010279">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010279</a></p> <p>Release Date: 2019-07-18</p> <p>Fix Resolution: 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in suricatasuricata cve high severity vulnerability vulnerable library suricatasuricata suricata git repository maintained by the oisf library home page a href found in head commit a href found in base branch master vulnerable source files learnage original sources party suricata tar suricata src detect c vulnerability details open information security foundation suricata prior to version is affected by denial of service tcp http detection bypass the impact is an attacker can evade a signature detection with a specialy formed sequence of network packets the component is detect c the attack vector is an attacker can trigger the vulnerability by a specifically crafted network tcp session the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
829,854
31,926,532,965
IssuesEvent
2023-09-19 02:30:23
kubestellar/kubestellar
https://api.github.com/repos/kubestellar/kubestellar
closed
feature: Adding support to select a subset of resources in a namespace to be synced to a workload execution cluster
kind/feature priority/important-soon partner
### Feature Description KubeStellar needs to support the use of EdgePlacement object to specify a subset of resources in a namespace to be synced to a particular workload execution cluster. This feature is required for some applications such as the following: - KubeFlow: the entire pipeline needs to run in one namespace at the same time, but we may want to run parts of the namespace in different clusters - MCAD: in some use-cases MCAD needs to cherry pick objects from a namespace to be shipped to a particular location. ### Proposed Solution KubeStellar EdgePlacement object allows selection of a subset of resources in a namespace to be synced to a particular workload execution cluster. ### Want to contribute? - [ ] I would like to work on this issue. ### Additional Context _No response_
1.0
feature: Adding support to select a subset of resources in a namespace to be synced to a workload execution cluster - ### Feature Description KubeStellar needs to support the use of EdgePlacement object to specify a subset of resources in a namespace to be synced to a particular workload execution cluster. This feature is required for some applications such as the following: - KubeFlow: the entire pipeline needs to run in one namespace at the same time, but we may want to run parts of the namespace in different clusters - MCAD: in some use-cases MCAD needs to cherry pick objects from a namespace to be shipped to a particular location. ### Proposed Solution KubeStellar EdgePlacement object allows selection of a subset of resources in a namespace to be synced to a particular workload execution cluster. ### Want to contribute? - [ ] I would like to work on this issue. ### Additional Context _No response_
non_infrastructure
feature adding support to select a subset of resources in a namespace to be synced to a workload execution cluster feature description kubestellar needs to support the use of edgeplacement object to specify a subset of resources in a namespace to be synced to a particular workload execution cluster this feature is required for some applications such as the following kubeflow the entire pipeline needs to run in one namespace at the same time but we may want to run parts of the namespace in different clusters mcad in some use cases mcad needs to cherry pick objects from a namespace to be shipped to a particular location proposed solution kubestellar edgeplacement object allows selection of a subset of resources in a namespace to be synced to a particular workload execution cluster want to contribute i would like to work on this issue additional context no response
0
143,585
22,094,067,187
IssuesEvent
2022-06-01 08:33:27
Eacaw/Inferno-GJ22
https://api.github.com/repos/Eacaw/Inferno-GJ22
opened
Fix up colliders
Enhancement Prefab Level Design
Get the game to a point where the player can't just force themselves through a wall. Remove the troves of superfluous invisible walls on some levels -_-. And Tidy everything up so that it all works nicely together. Intensive manual testing required by all parties.
1.0
Fix up colliders - Get the game to a point where the player can't just force themselves through a wall. Remove the troves of superfluous invisible walls on some levels -_-. And Tidy everything up so that it all works nicely together. Intensive manual testing required by all parties.
non_infrastructure
fix up colliders get the game to a point where the player can t just force themselves through a wall remove the troves of superfluous invisible walls on some levels and tidy everything up so that it all works nicely together intensive manual testing required by all parties
0
22,889
15,602,950,563
IssuesEvent
2021-03-19 00:53:19
etcd-io/website
https://api.github.com/repos/etcd-io/website
closed
Use latest version of Hugo to build the site
infrastructure
As of the time of creating this issue, the latest [release](https://github.com/gohugoio/hugo/releases) is 0.81.0. The currently used version, 0.65.0, is too old for docsy@HEAD. cc @nate-double-u -- let's discuss an upgrade strategy.
1.0
Use latest version of Hugo to build the site - As of the time of creating this issue, the latest [release](https://github.com/gohugoio/hugo/releases) is 0.81.0. The currently used version, 0.65.0, is too old for docsy@HEAD. cc @nate-double-u -- let's discuss an upgrade strategy.
infrastructure
use latest version of hugo to build the site as of the time of creating this issue the latest is the currently used version is too old for docsy head cc nate double u let s discuss an upgrade strategy
1
31,829
15,098,046,979
IssuesEvent
2021-02-07 21:05:08
mratsim/constantine
https://api.github.com/repos/mratsim/constantine
opened
BN254-Snarks: bad performance on Fp4 squaring
bug :beetle: performance :checkered_flag:
For some reason (multiplication by its super large 9+i non-residue?) BN254-Snarks is very slow on Fp4 ![image](https://user-images.githubusercontent.com/22738317/107159292-3b3c9d00-698f-11eb-8759-288a4f4a9045.png) Suspect 1: - Multiplication by Non-Residue here: https://github.com/mratsim/constantine/blob/c4a2dee42d39285a60813446dbb408f34f997114/constantine/tower_field_extensions/tower_instantiation.nim#L118-L129 - Rewritten from: https://github.com/mratsim/constantine/blob/2c5e12d5f893c6af9b97fec17d572268de3bd899/constantine/tower_field_extensions/tower_instantiation.nim#L88-L94 Suspect 2: The general squaring in quadratic field:https://github.com/mratsim/constantine/blob/c4a2dee42d39285a60813446dbb408f34f997114/constantine/tower_field_extensions/extension_fields.nim#L477-L524 The basic expression is `(c0² + β c1², 2 c0 c1)` We can either minimize Mul/Squarings, requiring only 2 multiplications by rewriting to: - `r0 = (c0 + c1)(c0 + β c1) - c0c1 - β c0c1` - `r1 = 2 c0c1` or only use squarings by rewritting to - `r0 = a0² + β a1²` - `r1 = (a0 + a1)² - a0² - a1²` The first expression requires 2 multiplication by non-residue.
True
BN254-Snarks: bad performance on Fp4 squaring - For some reason (multiplication by its super large 9+i non-residue?) BN254-Snarks is very slow on Fp4 ![image](https://user-images.githubusercontent.com/22738317/107159292-3b3c9d00-698f-11eb-8759-288a4f4a9045.png) Suspect 1: - Multiplication by Non-Residue here: https://github.com/mratsim/constantine/blob/c4a2dee42d39285a60813446dbb408f34f997114/constantine/tower_field_extensions/tower_instantiation.nim#L118-L129 - Rewritten from: https://github.com/mratsim/constantine/blob/2c5e12d5f893c6af9b97fec17d572268de3bd899/constantine/tower_field_extensions/tower_instantiation.nim#L88-L94 Suspect 2: The general squaring in quadratic field:https://github.com/mratsim/constantine/blob/c4a2dee42d39285a60813446dbb408f34f997114/constantine/tower_field_extensions/extension_fields.nim#L477-L524 The basic expression is `(c0² + β c1², 2 c0 c1)` We can either minimize Mul/Squarings, requiring only 2 multiplications by rewriting to: - `r0 = (c0 + c1)(c0 + β c1) - c0c1 - β c0c1` - `r1 = 2 c0c1` or only use squarings by rewritting to - `r0 = a0² + β a1²` - `r1 = (a0 + a1)² - a0² - a1²` The first expression requires 2 multiplication by non-residue.
non_infrastructure
snarks bad performance on squaring for some reason multiplication by its super large i non residue snarks is very slow on suspect multiplication by non residue here rewritten from suspect the general squaring in quadratic field the basic expression is β we can either minimize mul squarings requiring only multiplications by rewriting to β β or only use squarings by rewritting to β ² the first expression requires multiplication by non residue
0
15,510
11,569,500,736
IssuesEvent
2020-02-20 17:40:15
enarx/enarx
https://api.github.com/repos/enarx/enarx
opened
Test for SPDX ID headers in assembly files
infrastructure
This should be a fairly straightforward adaptation of the current SPDX ID test; it just needs to check for headers in `.c` and `.C` files in addition to whatever it's doing now.
1.0
Test for SPDX ID headers in assembly files - This should be a fairly straightforward adaptation of the current SPDX ID test; it just needs to check for headers in `.c` and `.C` files in addition to whatever it's doing now.
infrastructure
test for spdx id headers in assembly files this should be a fairly straightforward adaptation of the current spdx id test it just needs to check for headers in c and c files in addition to whatever it s doing now
1
301,101
22,716,032,999
IssuesEvent
2022-07-06 02:11:39
UnBArqDsw2022-1/2022_1_grupo5
https://api.github.com/repos/UnBArqDsw2022-1/2022_1_grupo5
opened
Diagrama de tecnologias
documentation
### Contact Details (optional) _No response_ ### Summary *** Documento contendo diagrama das tecnologias usadas: FrontEnd, API's, BackEnd (+ possíveis alternativas/intenções de uso) *** [SPRINT 2] Diagrama de tecnologias: - [ ] Definir todas as tecnologias a serem utilizadas no projeto. - [ ] Encaixá-las no diagrama BPWM - [ ] Acrescentar opções/alternativas de tecnologias(opcional). ### Motivation Estabelecer um protocolo de tecnologias a serem estudadas e utlizadas no ciclo de vida do projeto. ### Alternatives _No response_ ### Additional Context _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Diagrama de tecnologias - ### Contact Details (optional) _No response_ ### Summary *** Documento contendo diagrama das tecnologias usadas: FrontEnd, API's, BackEnd (+ possíveis alternativas/intenções de uso) *** [SPRINT 2] Diagrama de tecnologias: - [ ] Definir todas as tecnologias a serem utilizadas no projeto. - [ ] Encaixá-las no diagrama BPWM - [ ] Acrescentar opções/alternativas de tecnologias(opcional). ### Motivation Estabelecer um protocolo de tecnologias a serem estudadas e utlizadas no ciclo de vida do projeto. ### Alternatives _No response_ ### Additional Context _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_infrastructure
diagrama de tecnologias contact details optional no response summary documento contendo diagrama das tecnologias usadas frontend api s backend possíveis alternativas intenções de uso diagrama de tecnologias definir todas as tecnologias a serem utilizadas no projeto encaixá las no diagrama bpwm acrescentar opções alternativas de tecnologias opcional motivation estabelecer um protocolo de tecnologias a serem estudadas e utlizadas no ciclo de vida do projeto alternatives no response additional context no response code of conduct i agree to follow this project s code of conduct
0
28,582
23,364,453,182
IssuesEvent
2022-08-10 14:19:22
WordPress/performance
https://api.github.com/repos/WordPress/performance
closed
Remove plugin option for Multisite
[Type] Bug Infrastructure
## Bug Description Its follow-up enhancement for [#353](https://github.com/WordPress/performance/pull/345) and feedback https://github.com/WordPress/performance/pull/345#issuecomment-1155608644 If we set up multisite in WP then it will save the option in their sub-site DB. Like if you have two sub-site then the plugin option will save the option in the below tables. wp_options - Main site option table wp_2_options - 1st Sub-site option table wp_3_options - 2nd Sub-site option table The current uninstaller route only deletes options from the main site `perflab_modules_settings` but it will not remove the option for its subsite. ## Steps to reproduce <!-- Please provide detailed steps on how to reproduce the bug. --> 1. Setup [Multisite](https://wordpress.org/support/article/create-a-network/) 2. Create 2-3 subsite 3. Download and activate the Performance Lab plugin for all sites. 4. Delete plugin 5. Check the DB `perflab_modules_settings` option for sub-site. ## Screenshots ![image](https://user-images.githubusercontent.com/10103365/174580062-d62a6c72-7fa9-47fe-b341-5f3eacf9b184.png)
1.0
Remove plugin option for Multisite - ## Bug Description Its follow-up enhancement for [#353](https://github.com/WordPress/performance/pull/345) and feedback https://github.com/WordPress/performance/pull/345#issuecomment-1155608644 If we set up multisite in WP then it will save the option in their sub-site DB. Like if you have two sub-site then the plugin option will save the option in the below tables. wp_options - Main site option table wp_2_options - 1st Sub-site option table wp_3_options - 2nd Sub-site option table The current uninstaller route only deletes options from the main site `perflab_modules_settings` but it will not remove the option for its subsite. ## Steps to reproduce <!-- Please provide detailed steps on how to reproduce the bug. --> 1. Setup [Multisite](https://wordpress.org/support/article/create-a-network/) 2. Create 2-3 subsite 3. Download and activate the Performance Lab plugin for all sites. 4. Delete plugin 5. Check the DB `perflab_modules_settings` option for sub-site. ## Screenshots ![image](https://user-images.githubusercontent.com/10103365/174580062-d62a6c72-7fa9-47fe-b341-5f3eacf9b184.png)
infrastructure
remove plugin option for multisite bug description its follow up enhancement for and feedback if we set up multisite in wp then it will save the option in their sub site db like if you have two sub site then the plugin option will save the option in the below tables wp options main site option table wp options sub site option table wp options sub site option table the current uninstaller route only deletes options from the main site perflab modules settings but it will not remove the option for its subsite steps to reproduce setup create subsite download and activate the performance lab plugin for all sites delete plugin check the db perflab modules settings option for sub site screenshots
1
12,655
9,888,399,902
IssuesEvent
2019-06-25 11:28:18
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Set up database for MVP2
Epic data-storage infrastructure team-steam
Set up database @acn-dgopa we thought it could be a good idea to use an epic, need help filling it out
1.0
Set up database for MVP2 - Set up database @acn-dgopa we thought it could be a good idea to use an epic, need help filling it out
infrastructure
set up database for set up database acn dgopa we thought it could be a good idea to use an epic need help filling it out
1
180,106
30,440,341,351
IssuesEvent
2023-07-15 02:02:00
opensearch-project/sql
https://api.github.com/repos/opensearch-project/sql
opened
[RFC] Logical Plan Optimizer rework
documentation enhancement design
Please, track first design draft and discussion in #1752 # TL;DR Current behavior: ```py def Optimize: for node in PlanTree: # Traverse the Logical Plan Tree for rule in rules: # Enumerate rules tryApplyRule() ``` New behavior: ```py def Optimize: for rule in rules: # Enumerate rules for node in PlanTree: # Traverse the Logical Plan Tree tryApplyRule() ``` No new features, all tests pass, nothing changed for the end-user. # Background Currently each storage engine adds its own logical operator as concrete implementation for `TableScanOperator` abstraction. Typically each data source needs to add 2 logical operators for table scan with without aggregation. Take OpenSearch for example, there are `OpenSearchLogicalIndexScan` and `OpenSearchLogicalIndexAgg` and a bunch of pushdown optimization rules for each accordingly. ```py class LogicalPlanOptimizer: /* * OpenSearch rules include: * PUSH_DOWN_PAGE_SIZE * PUSH_DOWN_FILTER * PUSH_DOWN_AGGREGATION * PUSH_DOWN_SORT * PUSH_DOWN_HIGHLIGHT * PUSH_DOWN_NESTED * PUSH_DOWN_PROJECT * PUSH_DOWN_LIMIT * * that return *OpenSearchLogicalIndexAgg* * or *OpenSearchLogicalIndexScan* finally */ val rules: List<Rule> def optimize(plan: LogicalPlan): for rule in rules: # Enumerate rules for node in plan: # Traverse the Logical Plan Tree tryApplyRule() ``` # Optimization Protocol There are optimizaion guidelines which should be strictly followed to ensure that search query built completely matches user request. ## **1.** `Optimizer` should apply rules in the strict order they are defined. For example, `PUSH_DOWN_LIMIT` should be applied last. Violation of that causes bugs, for example [#1764](https://github.com/opensearch-project/sql/issues/1764), [#1774](https://github.com/opensearch-project/sql/issues/1774), [#1788](https://github.com/opensearch-project/sql/issues/1788). Sample queries: 1. Pagination with `LIMIT` ```json { "query": "SELECT * from `calcs` limit 10", "fetch_size": 3 } ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalPaginate" as Paginate state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalRelation" as Relation Paginate --> ProjectB ProjectB --> LimitB LimitB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` 2. `NESTED` with `LIMIT` ```sql SELECT nested(message.*) from nested-type limit 4 ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalNested" as Nested state "LogicalLimit" as LimitB state "LogicalRelation" as Relation ProjectB --> Nested Nested --> LimitB LimitB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` ## **2.** Optimizer should be able to apply a rule matching `Something` even when plan tree has something in between of `LogicalSomething` and `TableScanBuilder`, unless exception specified. `TableScanBuilder` could be wrapped by another tree node, for example in join implementation (see [#1623](https://github.com/opensearch-project/sql/pull/1623)). Sample queries: 1. `PPL` with `SORT` then `LIMIT` ```sql source=calcs | sort - int0 | head 10 | fields int0; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort Sort --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 2. `PPL` with `LIMIT` then `SORT` ```sql source=calcs | head 10 | sort - int0 | fields int0; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalSort" as Sort state "LogicalLimit" as Limit state "LogicalRelation" as Relation ProjectB --> Sort Sort --> Limit Limit --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 3. An `SQL` query; likely plan trees of all SQL queries are always in the same order ```sql select * from calcs where int0 > 0 order by int2 limit 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort Sort --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **3.** A query might have multiple `highlight`s backed by `LogicalHighlight` (and `filter`s and `sort`s) - all of them should be pushed down. A corresponding rule should be attempted multiple times. Sample queries: 1. Multiple `highlight`s ```sql SELECT highlight(Title), highlight(Body, pre_tags='<mark style="background-color: green;">', post_tags='</mark>') FROM beer.stackexchange WHERE multi_match([Title, Body], 'IPA') ORDER BY Id LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalHighlight" as Highlight1 state "LogicalHighlight" as Highlight2 state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Highlight1 Highlight1 --> Highlight2 Highlight2 --> Limit Limit --> Sort Sort --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 2. Multiple `filter`s ```sql source=account | where age > 30 | where age < 35 | fields age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalFilter" as Filter2 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Filter2 Filter2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 3. Multiple `sort`s ```sql source=account | sort age | sort lastname | head 20; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort1 state "LogicalSort" as Sort2 state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort1 Sort1 --> Sort2 Sort2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **4.** Dislike **3** most of the rules should be applied once only. **TODO** :exclamation: optimization doesn't work correctly in that cases Sample queries: 1. `PPL` ```sql source=account | fields firstname, lastname | head 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as LimitB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> LimitB LimitB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> Project2A Project2A --> TableScanBuilder } ``` ## **5.** `PUSH_DOWN_PROJECT` should not happen if there is a `LogicalEval` between `LogicalProject` and `TableScanBuilder`. Sample queries: ```sql source=bank | eval f = abs(age) | fields f; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalEval" as EvalB state "LogicalRelation" as Relation ProjectB --> EvalB EvalB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalEval" as EvalA state "TableScanBuilder" as TableScanBuilder ProjectA --> EvalA EvalA --> TableScanBuilder } ``` ## **6.** Similar to **5**, `LogicalWindow` in the plan tree between `LogicalProject` and `TableScanBuilder` should block `PUSH_DOWN_PROJECT` operation. Sample queries: ```sql SELECT avg(date0) OVER(PARTITION BY datetime1) from calcs; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalWindow" as WindowB state "LogicalSort" as Sort state "LogicalRelation" as Relation ProjectB --> WindowB WindowB --> Sort Sort --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalWindow" as WindowA state "TableScanBuilder" as TableScanBuilder ProjectA --> WindowA WindowA --> TableScanBuilder } ``` ## **7.** Some push down operations could be rejected (e.g. `pushDownWhatever` returns false), so corresponding `LogicalSomething` node remains in the tree. Avoid infinite re-trying to apply a rule for that node. Sample queries: 1. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownFilter` ```sql SELECT SUM(1) AS `cnt_` FROM calcs HAVING COUNT(2) > 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` 2. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownSort` and `pushDownLimit` ```sql SELECT COUNT(*) FROM account GROUP BY age ORDER BY COUNT(*) LIMIT 5; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalSort" as SortB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> SortB SortB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> SortA SortA --> TableScanBuilder } ``` ## **8.** `PUSH_DOWN_LIMIT` should be blocked if there is a `LogicalSort` or `LogicalFilter` between `LogicalLimit` and `TableScanBuilder`. Sample queries: 1. `PUSH_DOWN_SORT` can't be performed due to implementation restrictions ([#1471](https://github.com/opensearch-project/sql/issues/1471)), so `PUSH_DOWN_LIMIT` shouldn't be performed too. ```sql SELECT CAST(balance AS FLOAT) AS jdbc_float_alias FROM account ORDER BY jdbc_float_alias LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalSort" as SortB state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> SortB SortB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> SortA SortA --> TableScanBuilder } ``` 2. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownFilter`, so `LogicalFilter` remains in the tree and `PUSH_DOWN_LIMIT` shouldn't be performed. ```sql SELECT gender from account GROUP BY gender HAVING count(*) > 5 LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> FilterA FilterA --> TableScanBuilder } ``` ## **9.** `PUSH_DOWN_SORT` and `PUSH_DOWN_FILTER` should be after `PUSH_DOWN_AGGREGATION` if `LogicalSort` or `LogicalFilter` are on top of `LogicalAggregation`, see rules in `OpenSearchIndexScanAggregationBuilder`. Sample queries: 1. `SQL`: `LogicalFilter` is on top of `LogicalAggregation` ```sql SELECT gender from account GROUP BY gender HAVING count(*) > 500; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` 2. `PPL`: `LogicalFilter` is on top of `LogicalAggregation` ```sql source=account | stats sum(balance) as a by state | where a > 780000; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` ## **10.** `PUSH_DOWN_FILTER` should be before `PUSH_DOWN_AGGREGATION` if `LogicalFilteris` under of `LogicalAggregation`. Sample queries: 1. `LogicalFilter` is under `LogicalAggregation` ```sql SELECT gender from account WHERE age > 20 GROUP BY gender; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalAggregation" as Aggregation state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Aggregation Aggregation --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **11.** As combination of **9** and **10**, `PUSH_DOWN_FILTER` should be attempted before and after `PUSH_DOWN_AGGREGATION`. Sample queries: 1. `LogicalAggregation` surrounded by two `LogicalFilter`s ```sql SELECT gender from account WHERE age > 20 GROUP BY gender HAVING count(*) > 80; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1B state "LogicalAggregation" as Aggregation state "LogicalFilter" as Filter2B state "LogicalRelation" as Relation ProjectB --> Filter1B Filter1B --> Aggregation Aggregation --> Filter2B Filter2B --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` ## **12.** Subqueries: don’t `PUSH_DOWN_AGGREGATION`, `PUSH_DOWN_LIMIT`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_SORT` for the outer query, only for inner one (under most bottom `LogicalProject`). Sample queries: 1. Aggregation ```sql SELECT COUNT(*) FILTER(WHERE age > 35) FROM (SELECT * FROM bank) as a; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalAggregation" as AggergationB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> AggergationB AggergationB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalAggregation" as AggergationA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> AggergationA AggergationA --> Project2A Project2A --> TableScanBuilder } ``` 2. Filter ```sql SELECT origin FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights) AS f WHERE f.price > 100; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalFilter" as FilterB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> FilterB FilterB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalFilter" as FilterA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> FilterA FilterA --> Project2A Project2A --> TableScanBuilder } ``` 2. Sort ```sql SELECT origin FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights) AS f ORDER BY f.price; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalSort" as SortB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> SortB SortB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalSort" as SortA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> SortA SortA --> Project2A Project2A --> TableScanBuilder } ``` 3. Sort and Aggregation ```sql SELECT Origin, MIN(AvgTicketPrice) FROM (SELECT * FROM flights) AS flights GROUP BY Origin ORDER BY MAX(AvgTicketPrice) ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalSort" as SortB state "LogicalAggregation" as AggergationB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> SortB SortB --> AggergationB AggergationB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalSort" as SortA state "LogicalAggregation" as AggergationA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> SortA SortA --> AggergationA AggergationA --> Project2A Project2A --> TableScanBuilder } ``` 4. Sort and Limit ```sql SELECT price FROM (SELECT AvgTicketPrice AS price FROM flights LIMIT 10) AS flights ORDER BY price LIMIT 5, 5; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as Limit1B state "LogicalSort" as SortB state "LogicalProject" as Project2B state "LogicalLimit" as Limit2B state "LogicalRelation" as Relation Project1B --> Limit1B Limit1B --> SortB SortB --> Project2B Project2B --> Limit2B Limit2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> SortA SortA --> Project2A Project2A --> TableScanBuilder } ``` :exclamation: **TODO** add IT :exclamation: **TODO** nested, pagination, highlight, window ## **13.** Push down absolutely identical tree nodes (`LogicalSort`, `LogicalFilter`). Sample queries: 1. Sort ```sql SELECT FlightDelayMin, AvgTicketPrice, STDDEV_SAMP(AvgTicketPrice) OVER (ORDER BY FlightDelayMin) AS num FROM flights ORDER BY FlightDelayMin; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalWindow" as WindowB state "LogicalSort" as Sort1 state "LogicalSort" as Sort2 state "LogicalRelation" as Relation ProjectB --> WindowB WindowB --> Sort1 Sort1 --> Sort2 Sort2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalWindow" as WindowA state "TableScanBuilder" as TableScanBuilder ProjectA --> WindowA WindowA --> TableScanBuilder } ``` 2. Filter ```sql source=account | where age > 38 | where age > 38 | fields firstname, age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalFilter" as Filter2 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Filter2 Filter2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` :exclamation: **TBD** merge them (remove duplicates) or push down (current behavior - push down)? :exclamation: **TODO** add ITs (includehighlight) ## **14.** Optimize subqueries for `SQL` and complex queries in `PPL`. :exclamation: **TODO** not implemented. Sample queries: 1. Example from **4** - not optimized: ```sql source=account | fields firstname, lastname | head 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as LimitB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> LimitB LimitB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> Project2A Project2A --> TableScanBuilder } ``` 2. Relevance search with subquery: ```sql SELECT *, highlight(origin), _score FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights WHERE AvgTicketPrice > 100) AS f WHERE score(origin = match_query('Base')); ``` This query fails because `LogicalFilter` from outer query wasn’t pushed down, so V2 tried to do apply relevance search in memory. ``` UnsupportedOperationException: OpenSearch defined function [match_query] is only supported in WHERE and HAVING clause. ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalHighlight" as Highlight state "LogicalFilter" as Filter1B state "LogicalProject" as Project2B state "LogicalFilter" as Filter2B state "LogicalRelation" as Relation Project1B --> Highlight Highlight --> Filter1B Filter1B --> Project2B Project2B --> Filter2B Filter2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalFilter" as FilterA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> FilterA FilterA --> Project2A Project2A --> TableScanBuilder } ``` 3. Complex PPL query returns incorrect results ```sql source=account | where age > 30 | head 1000 | sort +balance | where age < 40 | head 100 | sort -balance | where balance > 10000 | fields age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalSort" as Sort1 state "LogicalLimit" as Limit1 state "LogicalFilter" as Filter2 state "LogicalSort" as Sort2 state "LogicalLimit" as Limit2 state "LogicalFilter" as Filter3 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Sort1 Sort1 --> Limit1 Limit1 --> Filter2 Filter2 --> Sort2 Sort2 --> Limit2 Limit2 --> Filter3 Filter3 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` # Optimizer rule update To satisfy requirement listed in **2**, new format rule format was created. See example for `PUSH_DOWN_FILTER` below; `...` matches any amount of Logical Plan Tree nodes of any types, except `LogicalFilter`. ```mermaid stateDiagram-v2 state "Old PUSH_DOWN_FILTER implementation" as OldFilter { state "LogicalFilter" as FilterOld state "ScanBuilder" as RelationOld FilterOld --> RelationOld } state "New PUSH_DOWN_FILTER implementation" as NewFilter { state "LogicalFilter" as FilterNew state "..." as dots state "ScanBuilder" as RelationNew FilterNew --> dots dots --> RelationNew } ``` This new format was applied to `PUSH_DOWN_PAGE_SIZE`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_AGGREGATION`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_SORT`, `PUSH_DOWN_HIGHLIGHT`, `PUSH_DOWN_NESTED`, `PUSH_DOWN_PROJECT` and `PUSH_DOWN_LIMIT`. `CreateTableScanBuilder`, `CreateTableWriteBuilder` and Prometheum related rules are not changed. ```mermaid stateDiagram-v2 state "CreateTableScanBuilder" as Builder { state "LogicalRelation" as Scan } ``` `PushDownRule` class used to build `PUSH_DOWN_*` rules. The class architecture follows: ```mermaid classDiagram class PushDownRule~T~ { -Class~T~ clazz -BiFunction pushDownFunction -List~Function~ exceptions -boolean canBeAppliedMultipleTimes* -getDefaultException() Function +PushDownRule(Class~T~, boolean, BiFunction) +PushDownRule(Class~T~, boolean, BiFunction, Function) +pattern()* Pattern~T~ +apply(T, Captures)* LogicalPlan -findTableScanBuilder(LogicalPlan) Optional~TableScanBuilder~ } ``` The following rule configurations are created: | Rule | Tree Node Type | Can be applied multiple times | Push Down function | Exception | |------|----------------|----------------------------------|-----------------------|------------------------------------------------| | `PUSH_DOWN_FILTER` | `LogicalFilter` | `true` | `pushDownFilter` | `LogicalAggregation`, `LogicalProject` | | `PUSH_DOWN_AGGREGATION` | `LogicalAggregation` | `false` | `pushDownAggregation` | `LogicalProject` | | `PUSH_DOWN_SORT` | `LogicalSort` | `true` | `pushDownSort` | `LogicalProject` | | `PUSH_DOWN_LIMIT` | `LogicalLimit` | `false` | `pushDownLimit` | `LogicalSort`, `LogicalFilter`, `LogicalProject` | | `PUSH_DOWN_PROJECT` | `LogicalProject` | `false` | `pushDownProject` | `LogicalEval`, `LogicalWindow` | | `PUSH_DOWN_HIGHLIGHT` | `LogicalHighlight` | `true` | `pushDownHighlight` | none | | `PUSH_DOWN_NESTED` | `LogicalNested` | `false` | `pushDownNested` | none | | `PUSH_DOWN_PAGE_SIZE` | `LogicalPaginate` | `false` | `pushDownPageSize` | none | Note that default exception is applied to all rules. Rules are defined, checked and applied in the following order: ``` PUSH_DOWN_PAGE_SIZE PUSH_DOWN_FILTER PUSH_DOWN_AGGREGATION PUSH_DOWN_FILTER PUSH_DOWN_SORT PUSH_DOWN_HIGHLIGHT PUSH_DOWN_NESTED PUSH_DOWN_PROJECT PUSH_DOWN_LIMIT ``` To satisfy requirement **11**, `PUSH_DOWN_FILTER` is listed twice. # Code Please, see code MVP PoC in `Bit-Quill:dev-optimizer-rework` branch: https://github.com/opensearch-project/sql/compare/main..Bit-Quill:opensearch-project-sql:dev-optimizer-rework?expand=1 This code passes all IT and contains only Optimizer rework. There're no new features, so things marked as not implemented are still not implemented. # References Current optimizer doc: https://github.com/opensearch-project/sql/blob/main/docs/dev/query-optimizer-improvement.md # Purpose Proposed changes will unblock other fixes and features: - [ ] #1752 - [ ] #1764 - [ ] #1774 - [ ] #1788 - [ ] #494
1.0
[RFC] Logical Plan Optimizer rework - Please, track first design draft and discussion in #1752 # TL;DR Current behavior: ```py def Optimize: for node in PlanTree: # Traverse the Logical Plan Tree for rule in rules: # Enumerate rules tryApplyRule() ``` New behavior: ```py def Optimize: for rule in rules: # Enumerate rules for node in PlanTree: # Traverse the Logical Plan Tree tryApplyRule() ``` No new features, all tests pass, nothing changed for the end-user. # Background Currently each storage engine adds its own logical operator as concrete implementation for `TableScanOperator` abstraction. Typically each data source needs to add 2 logical operators for table scan with without aggregation. Take OpenSearch for example, there are `OpenSearchLogicalIndexScan` and `OpenSearchLogicalIndexAgg` and a bunch of pushdown optimization rules for each accordingly. ```py class LogicalPlanOptimizer: /* * OpenSearch rules include: * PUSH_DOWN_PAGE_SIZE * PUSH_DOWN_FILTER * PUSH_DOWN_AGGREGATION * PUSH_DOWN_SORT * PUSH_DOWN_HIGHLIGHT * PUSH_DOWN_NESTED * PUSH_DOWN_PROJECT * PUSH_DOWN_LIMIT * * that return *OpenSearchLogicalIndexAgg* * or *OpenSearchLogicalIndexScan* finally */ val rules: List<Rule> def optimize(plan: LogicalPlan): for rule in rules: # Enumerate rules for node in plan: # Traverse the Logical Plan Tree tryApplyRule() ``` # Optimization Protocol There are optimizaion guidelines which should be strictly followed to ensure that search query built completely matches user request. ## **1.** `Optimizer` should apply rules in the strict order they are defined. For example, `PUSH_DOWN_LIMIT` should be applied last. Violation of that causes bugs, for example [#1764](https://github.com/opensearch-project/sql/issues/1764), [#1774](https://github.com/opensearch-project/sql/issues/1774), [#1788](https://github.com/opensearch-project/sql/issues/1788). Sample queries: 1. Pagination with `LIMIT` ```json { "query": "SELECT * from `calcs` limit 10", "fetch_size": 3 } ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalPaginate" as Paginate state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalRelation" as Relation Paginate --> ProjectB ProjectB --> LimitB LimitB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` 2. `NESTED` with `LIMIT` ```sql SELECT nested(message.*) from nested-type limit 4 ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalNested" as Nested state "LogicalLimit" as LimitB state "LogicalRelation" as Relation ProjectB --> Nested Nested --> LimitB LimitB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` ## **2.** Optimizer should be able to apply a rule matching `Something` even when plan tree has something in between of `LogicalSomething` and `TableScanBuilder`, unless exception specified. `TableScanBuilder` could be wrapped by another tree node, for example in join implementation (see [#1623](https://github.com/opensearch-project/sql/pull/1623)). Sample queries: 1. `PPL` with `SORT` then `LIMIT` ```sql source=calcs | sort - int0 | head 10 | fields int0; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort Sort --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 2. `PPL` with `LIMIT` then `SORT` ```sql source=calcs | head 10 | sort - int0 | fields int0; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalSort" as Sort state "LogicalLimit" as Limit state "LogicalRelation" as Relation ProjectB --> Sort Sort --> Limit Limit --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 3. An `SQL` query; likely plan trees of all SQL queries are always in the same order ```sql select * from calcs where int0 > 0 order by int2 limit 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort Sort --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **3.** A query might have multiple `highlight`s backed by `LogicalHighlight` (and `filter`s and `sort`s) - all of them should be pushed down. A corresponding rule should be attempted multiple times. Sample queries: 1. Multiple `highlight`s ```sql SELECT highlight(Title), highlight(Body, pre_tags='<mark style="background-color: green;">', post_tags='</mark>') FROM beer.stackexchange WHERE multi_match([Title, Body], 'IPA') ORDER BY Id LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalHighlight" as Highlight1 state "LogicalHighlight" as Highlight2 state "LogicalLimit" as Limit state "LogicalSort" as Sort state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Highlight1 Highlight1 --> Highlight2 Highlight2 --> Limit Limit --> Sort Sort --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 2. Multiple `filter`s ```sql source=account | where age > 30 | where age < 35 | fields age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalFilter" as Filter2 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Filter2 Filter2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` 3. Multiple `sort`s ```sql source=account | sort age | sort lastname | head 20; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as Limit state "LogicalSort" as Sort1 state "LogicalSort" as Sort2 state "LogicalRelation" as Relation ProjectB --> Limit Limit --> Sort1 Sort1 --> Sort2 Sort2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **4.** Dislike **3** most of the rules should be applied once only. **TODO** :exclamation: optimization doesn't work correctly in that cases Sample queries: 1. `PPL` ```sql source=account | fields firstname, lastname | head 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as LimitB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> LimitB LimitB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> Project2A Project2A --> TableScanBuilder } ``` ## **5.** `PUSH_DOWN_PROJECT` should not happen if there is a `LogicalEval` between `LogicalProject` and `TableScanBuilder`. Sample queries: ```sql source=bank | eval f = abs(age) | fields f; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalEval" as EvalB state "LogicalRelation" as Relation ProjectB --> EvalB EvalB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalEval" as EvalA state "TableScanBuilder" as TableScanBuilder ProjectA --> EvalA EvalA --> TableScanBuilder } ``` ## **6.** Similar to **5**, `LogicalWindow` in the plan tree between `LogicalProject` and `TableScanBuilder` should block `PUSH_DOWN_PROJECT` operation. Sample queries: ```sql SELECT avg(date0) OVER(PARTITION BY datetime1) from calcs; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalWindow" as WindowB state "LogicalSort" as Sort state "LogicalRelation" as Relation ProjectB --> WindowB WindowB --> Sort Sort --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalWindow" as WindowA state "TableScanBuilder" as TableScanBuilder ProjectA --> WindowA WindowA --> TableScanBuilder } ``` ## **7.** Some push down operations could be rejected (e.g. `pushDownWhatever` returns false), so corresponding `LogicalSomething` node remains in the tree. Avoid infinite re-trying to apply a rule for that node. Sample queries: 1. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownFilter` ```sql SELECT SUM(1) AS `cnt_` FROM calcs HAVING COUNT(2) > 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` 2. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownSort` and `pushDownLimit` ```sql SELECT COUNT(*) FROM account GROUP BY age ORDER BY COUNT(*) LIMIT 5; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalSort" as SortB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> SortB SortB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> SortA SortA --> TableScanBuilder } ``` ## **8.** `PUSH_DOWN_LIMIT` should be blocked if there is a `LogicalSort` or `LogicalFilter` between `LogicalLimit` and `TableScanBuilder`. Sample queries: 1. `PUSH_DOWN_SORT` can't be performed due to implementation restrictions ([#1471](https://github.com/opensearch-project/sql/issues/1471)), so `PUSH_DOWN_LIMIT` shouldn't be performed too. ```sql SELECT CAST(balance AS FLOAT) AS jdbc_float_alias FROM account ORDER BY jdbc_float_alias LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalSort" as SortB state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> SortB SortB --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> SortA SortA --> TableScanBuilder } ``` 2. `OpenSearchIndexScanAggregationBuilder` rejects `pushDownFilter`, so `LogicalFilter` remains in the tree and `PUSH_DOWN_LIMIT` shouldn't be performed. ```sql SELECT gender from account GROUP BY gender HAVING count(*) > 5 LIMIT 1; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalLimit" as LimitB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> LimitB LimitB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> FilterA FilterA --> TableScanBuilder } ``` ## **9.** `PUSH_DOWN_SORT` and `PUSH_DOWN_FILTER` should be after `PUSH_DOWN_AGGREGATION` if `LogicalSort` or `LogicalFilter` are on top of `LogicalAggregation`, see rules in `OpenSearchIndexScanAggregationBuilder`. Sample queries: 1. `SQL`: `LogicalFilter` is on top of `LogicalAggregation` ```sql SELECT gender from account GROUP BY gender HAVING count(*) > 500; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` 2. `PPL`: `LogicalFilter` is on top of `LogicalAggregation` ```sql source=account | stats sum(balance) as a by state | where a > 780000; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as FilterB state "LogicalAggregation" as Aggregation state "LogicalRelation" as Relation ProjectB --> FilterB FilterB --> Aggregation Aggregation --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` ## **10.** `PUSH_DOWN_FILTER` should be before `PUSH_DOWN_AGGREGATION` if `LogicalFilteris` under of `LogicalAggregation`. Sample queries: 1. `LogicalFilter` is under `LogicalAggregation` ```sql SELECT gender from account WHERE age > 20 GROUP BY gender; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalAggregation" as Aggregation state "LogicalFilter" as Filter state "LogicalRelation" as Relation ProjectB --> Aggregation Aggregation --> Filter Filter --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` ## **11.** As combination of **9** and **10**, `PUSH_DOWN_FILTER` should be attempted before and after `PUSH_DOWN_AGGREGATION`. Sample queries: 1. `LogicalAggregation` surrounded by two `LogicalFilter`s ```sql SELECT gender from account WHERE age > 20 GROUP BY gender HAVING count(*) > 80; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1B state "LogicalAggregation" as Aggregation state "LogicalFilter" as Filter2B state "LogicalRelation" as Relation ProjectB --> Filter1B Filter1B --> Aggregation Aggregation --> Filter2B Filter2B --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalFilter" as FilterA state "TableScanBuilder" as TableScanBuilder ProjectA --> FilterA FilterA --> TableScanBuilder } ``` ## **12.** Subqueries: don’t `PUSH_DOWN_AGGREGATION`, `PUSH_DOWN_LIMIT`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_SORT` for the outer query, only for inner one (under most bottom `LogicalProject`). Sample queries: 1. Aggregation ```sql SELECT COUNT(*) FILTER(WHERE age > 35) FROM (SELECT * FROM bank) as a; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalAggregation" as AggergationB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> AggergationB AggergationB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalAggregation" as AggergationA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> AggergationA AggergationA --> Project2A Project2A --> TableScanBuilder } ``` 2. Filter ```sql SELECT origin FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights) AS f WHERE f.price > 100; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalFilter" as FilterB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> FilterB FilterB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalFilter" as FilterA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> FilterA FilterA --> Project2A Project2A --> TableScanBuilder } ``` 2. Sort ```sql SELECT origin FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights) AS f ORDER BY f.price; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalSort" as SortB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> SortB SortB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalSort" as SortA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> SortA SortA --> Project2A Project2A --> TableScanBuilder } ``` 3. Sort and Aggregation ```sql SELECT Origin, MIN(AvgTicketPrice) FROM (SELECT * FROM flights) AS flights GROUP BY Origin ORDER BY MAX(AvgTicketPrice) ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalSort" as SortB state "LogicalAggregation" as AggergationB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> SortB SortB --> AggergationB AggergationB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalSort" as SortA state "LogicalAggregation" as AggergationA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> SortA SortA --> AggergationA AggergationA --> Project2A Project2A --> TableScanBuilder } ``` 4. Sort and Limit ```sql SELECT price FROM (SELECT AvgTicketPrice AS price FROM flights LIMIT 10) AS flights ORDER BY price LIMIT 5, 5; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as Limit1B state "LogicalSort" as SortB state "LogicalProject" as Project2B state "LogicalLimit" as Limit2B state "LogicalRelation" as Relation Project1B --> Limit1B Limit1B --> SortB SortB --> Project2B Project2B --> Limit2B Limit2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalSort" as SortA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> SortA SortA --> Project2A Project2A --> TableScanBuilder } ``` :exclamation: **TODO** add IT :exclamation: **TODO** nested, pagination, highlight, window ## **13.** Push down absolutely identical tree nodes (`LogicalSort`, `LogicalFilter`). Sample queries: 1. Sort ```sql SELECT FlightDelayMin, AvgTicketPrice, STDDEV_SAMP(AvgTicketPrice) OVER (ORDER BY FlightDelayMin) AS num FROM flights ORDER BY FlightDelayMin; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalWindow" as WindowB state "LogicalSort" as Sort1 state "LogicalSort" as Sort2 state "LogicalRelation" as Relation ProjectB --> WindowB WindowB --> Sort1 Sort1 --> Sort2 Sort2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalWindow" as WindowA state "TableScanBuilder" as TableScanBuilder ProjectA --> WindowA WindowA --> TableScanBuilder } ``` 2. Filter ```sql source=account | where age > 38 | where age > 38 | fields firstname, age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalFilter" as Filter2 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Filter2 Filter2 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "TableScanBuilder" as TableScanBuilder ProjectA --> TableScanBuilder } ``` :exclamation: **TBD** merge them (remove duplicates) or push down (current behavior - push down)? :exclamation: **TODO** add ITs (includehighlight) ## **14.** Optimize subqueries for `SQL` and complex queries in `PPL`. :exclamation: **TODO** not implemented. Sample queries: 1. Example from **4** - not optimized: ```sql source=account | fields firstname, lastname | head 10; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalLimit" as LimitB state "LogicalProject" as Project2B state "LogicalRelation" as Relation Project1B --> LimitB LimitB --> Project2B Project2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalLimit" as LimitA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> LimitA LimitA --> Project2A Project2A --> TableScanBuilder } ``` 2. Relevance search with subquery: ```sql SELECT *, highlight(origin), _score FROM (SELECT Origin AS origin, AvgTicketPrice AS price FROM flights WHERE AvgTicketPrice > 100) AS f WHERE score(origin = match_query('Base')); ``` This query fails because `LogicalFilter` from outer query wasn’t pushed down, so V2 tried to do apply relevance search in memory. ``` UnsupportedOperationException: OpenSearch defined function [match_query] is only supported in WHERE and HAVING clause. ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as Project1B state "LogicalHighlight" as Highlight state "LogicalFilter" as Filter1B state "LogicalProject" as Project2B state "LogicalFilter" as Filter2B state "LogicalRelation" as Relation Project1B --> Highlight Highlight --> Filter1B Filter1B --> Project2B Project2B --> Filter2B Filter2B --> Relation } state "After" as After { state "LogicalProject" as Project1A state "LogicalFilter" as FilterA state "LogicalProject" as Project2A state "TableScanBuilder" as TableScanBuilder Project1A --> FilterA FilterA --> Project2A Project2A --> TableScanBuilder } ``` 3. Complex PPL query returns incorrect results ```sql source=account | where age > 30 | head 1000 | sort +balance | where age < 40 | head 100 | sort -balance | where balance > 10000 | fields age; ``` ```mermaid stateDiagram-v2 state "Before" as Before { state "LogicalProject" as ProjectB state "LogicalFilter" as Filter1 state "LogicalSort" as Sort1 state "LogicalLimit" as Limit1 state "LogicalFilter" as Filter2 state "LogicalSort" as Sort2 state "LogicalLimit" as Limit2 state "LogicalFilter" as Filter3 state "LogicalRelation" as Relation ProjectB --> Filter1 Filter1 --> Sort1 Sort1 --> Limit1 Limit1 --> Filter2 Filter2 --> Sort2 Sort2 --> Limit2 Limit2 --> Filter3 Filter3 --> Relation } state "After" as After { state "LogicalProject" as ProjectA state "LogicalLimit" as LimitA state "TableScanBuilder" as TableScanBuilder ProjectA --> LimitA LimitA --> TableScanBuilder } ``` # Optimizer rule update To satisfy requirement listed in **2**, new format rule format was created. See example for `PUSH_DOWN_FILTER` below; `...` matches any amount of Logical Plan Tree nodes of any types, except `LogicalFilter`. ```mermaid stateDiagram-v2 state "Old PUSH_DOWN_FILTER implementation" as OldFilter { state "LogicalFilter" as FilterOld state "ScanBuilder" as RelationOld FilterOld --> RelationOld } state "New PUSH_DOWN_FILTER implementation" as NewFilter { state "LogicalFilter" as FilterNew state "..." as dots state "ScanBuilder" as RelationNew FilterNew --> dots dots --> RelationNew } ``` This new format was applied to `PUSH_DOWN_PAGE_SIZE`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_AGGREGATION`, `PUSH_DOWN_FILTER`, `PUSH_DOWN_SORT`, `PUSH_DOWN_HIGHLIGHT`, `PUSH_DOWN_NESTED`, `PUSH_DOWN_PROJECT` and `PUSH_DOWN_LIMIT`. `CreateTableScanBuilder`, `CreateTableWriteBuilder` and Prometheum related rules are not changed. ```mermaid stateDiagram-v2 state "CreateTableScanBuilder" as Builder { state "LogicalRelation" as Scan } ``` `PushDownRule` class used to build `PUSH_DOWN_*` rules. The class architecture follows: ```mermaid classDiagram class PushDownRule~T~ { -Class~T~ clazz -BiFunction pushDownFunction -List~Function~ exceptions -boolean canBeAppliedMultipleTimes* -getDefaultException() Function +PushDownRule(Class~T~, boolean, BiFunction) +PushDownRule(Class~T~, boolean, BiFunction, Function) +pattern()* Pattern~T~ +apply(T, Captures)* LogicalPlan -findTableScanBuilder(LogicalPlan) Optional~TableScanBuilder~ } ``` The following rule configurations are created: | Rule | Tree Node Type | Can be applied multiple times | Push Down function | Exception | |------|----------------|----------------------------------|-----------------------|------------------------------------------------| | `PUSH_DOWN_FILTER` | `LogicalFilter` | `true` | `pushDownFilter` | `LogicalAggregation`, `LogicalProject` | | `PUSH_DOWN_AGGREGATION` | `LogicalAggregation` | `false` | `pushDownAggregation` | `LogicalProject` | | `PUSH_DOWN_SORT` | `LogicalSort` | `true` | `pushDownSort` | `LogicalProject` | | `PUSH_DOWN_LIMIT` | `LogicalLimit` | `false` | `pushDownLimit` | `LogicalSort`, `LogicalFilter`, `LogicalProject` | | `PUSH_DOWN_PROJECT` | `LogicalProject` | `false` | `pushDownProject` | `LogicalEval`, `LogicalWindow` | | `PUSH_DOWN_HIGHLIGHT` | `LogicalHighlight` | `true` | `pushDownHighlight` | none | | `PUSH_DOWN_NESTED` | `LogicalNested` | `false` | `pushDownNested` | none | | `PUSH_DOWN_PAGE_SIZE` | `LogicalPaginate` | `false` | `pushDownPageSize` | none | Note that default exception is applied to all rules. Rules are defined, checked and applied in the following order: ``` PUSH_DOWN_PAGE_SIZE PUSH_DOWN_FILTER PUSH_DOWN_AGGREGATION PUSH_DOWN_FILTER PUSH_DOWN_SORT PUSH_DOWN_HIGHLIGHT PUSH_DOWN_NESTED PUSH_DOWN_PROJECT PUSH_DOWN_LIMIT ``` To satisfy requirement **11**, `PUSH_DOWN_FILTER` is listed twice. # Code Please, see code MVP PoC in `Bit-Quill:dev-optimizer-rework` branch: https://github.com/opensearch-project/sql/compare/main..Bit-Quill:opensearch-project-sql:dev-optimizer-rework?expand=1 This code passes all IT and contains only Optimizer rework. There're no new features, so things marked as not implemented are still not implemented. # References Current optimizer doc: https://github.com/opensearch-project/sql/blob/main/docs/dev/query-optimizer-improvement.md # Purpose Proposed changes will unblock other fixes and features: - [ ] #1752 - [ ] #1764 - [ ] #1774 - [ ] #1788 - [ ] #494
non_infrastructure
logical plan optimizer rework please track first design draft and discussion in tl dr current behavior py def optimize for node in plantree traverse the logical plan tree for rule in rules enumerate rules tryapplyrule new behavior py def optimize for rule in rules enumerate rules for node in plantree traverse the logical plan tree tryapplyrule no new features all tests pass nothing changed for the end user background currently each storage engine adds its own logical operator as concrete implementation for tablescanoperator abstraction typically each data source needs to add logical operators for table scan with without aggregation take opensearch for example there are opensearchlogicalindexscan and opensearchlogicalindexagg and a bunch of pushdown optimization rules for each accordingly py class logicalplanoptimizer opensearch rules include push down page size push down filter push down aggregation push down sort push down highlight push down nested push down project push down limit that return opensearchlogicalindexagg or opensearchlogicalindexscan finally val rules list def optimize plan logicalplan for rule in rules enumerate rules for node in plan traverse the logical plan tree tryapplyrule optimization protocol there are optimizaion guidelines which should be strictly followed to ensure that search query built completely matches user request optimizer should apply rules in the strict order they are defined for example push down limit should be applied last violation of that causes bugs for example sample queries pagination with limit json query select from calcs limit fetch size mermaid statediagram state before as before state logicalpaginate as paginate state logicalproject as projectb state logicallimit as limitb state logicalrelation as relation paginate projectb projectb limitb limitb relation state after as after state logicalproject as projecta state logicallimit as limita state tablescanbuilder as tablescanbuilder projecta limita limita tablescanbuilder nested with limit sql select nested message from nested type limit mermaid statediagram state before as before state logicalproject as projectb state logicalnested as nested state logicallimit as limitb state logicalrelation as relation projectb nested nested limitb limitb relation state after as after state logicalproject as projecta state logicallimit as limita state tablescanbuilder as tablescanbuilder projecta limita limita tablescanbuilder optimizer should be able to apply a rule matching something even when plan tree has something in between of logicalsomething and tablescanbuilder unless exception specified tablescanbuilder could be wrapped by another tree node for example in join implementation see sample queries ppl with sort then limit sql source calcs sort head fields mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limit state logicalsort as sort state logicalrelation as relation projectb limit limit sort sort relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder ppl with limit then sort sql source calcs head sort fields mermaid statediagram state before as before state logicalproject as projectb state logicalsort as sort state logicallimit as limit state logicalrelation as relation projectb sort sort limit limit relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder an sql query likely plan trees of all sql queries are always in the same order sql select from calcs where order by limit mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limit state logicalsort as sort state logicalfilter as filter state logicalrelation as relation projectb limit limit sort sort filter filter relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder a query might have multiple highlight s backed by logicalhighlight and filter s and sort s all of them should be pushed down a corresponding rule should be attempted multiple times sample queries multiple highlight s sql select highlight title highlight body pre tags post tags from beer stackexchange where multi match ipa order by id limit mermaid statediagram state before as before state logicalproject as projectb state logicalhighlight as state logicalhighlight as state logicallimit as limit state logicalsort as sort state logicalfilter as filter state logicalrelation as relation projectb limit limit sort sort filter filter relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder multiple filter s sql source account where age where age fields age mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as state logicalfilter as state logicalrelation as relation projectb relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder multiple sort s sql source account sort age sort lastname head mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limit state logicalsort as state logicalsort as state logicalrelation as relation projectb limit limit relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder dislike most of the rules should be applied once only todo exclamation optimization doesn t work correctly in that cases sample queries ppl sql source account fields firstname lastname head mermaid statediagram state before as before state logicalproject as state logicallimit as limitb state logicalproject as state logicalrelation as relation limitb limitb relation state after as after state logicalproject as state logicallimit as limita state logicalproject as state tablescanbuilder as tablescanbuilder limita limita tablescanbuilder push down project should not happen if there is a logicaleval between logicalproject and tablescanbuilder sample queries sql source bank eval f abs age fields f mermaid statediagram state before as before state logicalproject as projectb state logicaleval as evalb state logicalrelation as relation projectb evalb evalb relation state after as after state logicalproject as projecta state logicaleval as evala state tablescanbuilder as tablescanbuilder projecta evala evala tablescanbuilder similar to logicalwindow in the plan tree between logicalproject and tablescanbuilder should block push down project operation sample queries sql select avg over partition by from calcs mermaid statediagram state before as before state logicalproject as projectb state logicalwindow as windowb state logicalsort as sort state logicalrelation as relation projectb windowb windowb sort sort relation state after as after state logicalproject as projecta state logicalwindow as windowa state tablescanbuilder as tablescanbuilder projecta windowa windowa tablescanbuilder some push down operations could be rejected e g pushdownwhatever returns false so corresponding logicalsomething node remains in the tree avoid infinite re trying to apply a rule for that node sample queries opensearchindexscanaggregationbuilder rejects pushdownfilter sql select sum as cnt from calcs having count mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as filterb state logicalaggregation as aggregation state logicalrelation as relation projectb filterb filterb aggregation aggregation relation state after as after state logicalproject as projecta state logicalfilter as filtera state tablescanbuilder as tablescanbuilder projecta filtera filtera tablescanbuilder opensearchindexscanaggregationbuilder rejects pushdownsort and pushdownlimit sql select count from account group by age order by count limit mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limitb state logicalsort as sortb state logicalaggregation as aggregation state logicalrelation as relation projectb limitb limitb sortb sortb aggregation aggregation relation state after as after state logicalproject as projecta state logicallimit as limita state logicalsort as sorta state tablescanbuilder as tablescanbuilder projecta limita limita sorta sorta tablescanbuilder push down limit should be blocked if there is a logicalsort or logicalfilter between logicallimit and tablescanbuilder sample queries push down sort can t be performed due to implementation restrictions so push down limit shouldn t be performed too sql select cast balance as float as jdbc float alias from account order by jdbc float alias limit mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limitb state logicalsort as sortb state logicalrelation as relation projectb limitb limitb sortb sortb relation state after as after state logicalproject as projecta state logicallimit as limita state logicalsort as sorta state tablescanbuilder as tablescanbuilder projecta limita limita sorta sorta tablescanbuilder opensearchindexscanaggregationbuilder rejects pushdownfilter so logicalfilter remains in the tree and push down limit shouldn t be performed sql select gender from account group by gender having count limit mermaid statediagram state before as before state logicalproject as projectb state logicallimit as limitb state logicalfilter as filterb state logicalaggregation as aggregation state logicalrelation as relation projectb limitb limitb filterb filterb aggregation aggregation relation state after as after state logicalproject as projecta state logicallimit as limita state logicalfilter as filtera state tablescanbuilder as tablescanbuilder projecta limita limita filtera filtera tablescanbuilder push down sort and push down filter should be after push down aggregation if logicalsort or logicalfilter are on top of logicalaggregation see rules in opensearchindexscanaggregationbuilder sample queries sql logicalfilter is on top of logicalaggregation sql select gender from account group by gender having count mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as filterb state logicalaggregation as aggregation state logicalrelation as relation projectb filterb filterb aggregation aggregation relation state after as after state logicalproject as projecta state logicalfilter as filtera state tablescanbuilder as tablescanbuilder projecta filtera filtera tablescanbuilder ppl logicalfilter is on top of logicalaggregation sql source account stats sum balance as a by state where a mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as filterb state logicalaggregation as aggregation state logicalrelation as relation projectb filterb filterb aggregation aggregation relation state after as after state logicalproject as projecta state logicalfilter as filtera state tablescanbuilder as tablescanbuilder projecta filtera filtera tablescanbuilder push down filter should be before push down aggregation if logicalfilteris under of logicalaggregation sample queries logicalfilter is under logicalaggregation sql select gender from account where age group by gender mermaid statediagram state before as before state logicalproject as projectb state logicalaggregation as aggregation state logicalfilter as filter state logicalrelation as relation projectb aggregation aggregation filter filter relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder as combination of and push down filter should be attempted before and after push down aggregation sample queries logicalaggregation surrounded by two logicalfilter s sql select gender from account where age group by gender having count mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as state logicalaggregation as aggregation state logicalfilter as state logicalrelation as relation projectb aggregation aggregation relation state after as after state logicalproject as projecta state logicalfilter as filtera state tablescanbuilder as tablescanbuilder projecta filtera filtera tablescanbuilder subqueries don’t push down aggregation push down limit push down filter push down sort for the outer query only for inner one under most bottom logicalproject sample queries aggregation sql select count filter where age from select from bank as a mermaid statediagram state before as before state logicalproject as state logicalaggregation as aggergationb state logicalproject as state logicalrelation as relation aggergationb aggergationb relation state after as after state logicalproject as state logicalaggregation as aggergationa state logicalproject as state tablescanbuilder as tablescanbuilder aggergationa aggergationa tablescanbuilder filter sql select origin from select origin as origin avgticketprice as price from flights as f where f price mermaid statediagram state before as before state logicalproject as state logicalfilter as filterb state logicalproject as state logicalrelation as relation filterb filterb relation state after as after state logicalproject as state logicalfilter as filtera state logicalproject as state tablescanbuilder as tablescanbuilder filtera filtera tablescanbuilder sort sql select origin from select origin as origin avgticketprice as price from flights as f order by f price mermaid statediagram state before as before state logicalproject as state logicalsort as sortb state logicalproject as state logicalrelation as relation sortb sortb relation state after as after state logicalproject as state logicalsort as sorta state logicalproject as state tablescanbuilder as tablescanbuilder sorta sorta tablescanbuilder sort and aggregation sql select origin min avgticketprice from select from flights as flights group by origin order by max avgticketprice mermaid statediagram state before as before state logicalproject as state logicalsort as sortb state logicalaggregation as aggergationb state logicalproject as state logicalrelation as relation sortb sortb aggergationb aggergationb relation state after as after state logicalproject as state logicalsort as sorta state logicalaggregation as aggergationa state logicalproject as state tablescanbuilder as tablescanbuilder sorta sorta aggergationa aggergationa tablescanbuilder sort and limit sql select price from select avgticketprice as price from flights limit as flights order by price limit mermaid statediagram state before as before state logicalproject as state logicallimit as state logicalsort as sortb state logicalproject as state logicallimit as state logicalrelation as relation sortb sortb relation state after as after state logicalproject as state logicallimit as limita state logicalsort as sorta state logicalproject as state tablescanbuilder as tablescanbuilder limita limita sorta sorta tablescanbuilder exclamation todo add it exclamation todo nested pagination highlight window push down absolutely identical tree nodes logicalsort logicalfilter sample queries sort sql select flightdelaymin avgticketprice stddev samp avgticketprice over order by flightdelaymin as num from flights order by flightdelaymin mermaid statediagram state before as before state logicalproject as projectb state logicalwindow as windowb state logicalsort as state logicalsort as state logicalrelation as relation projectb windowb windowb relation state after as after state logicalproject as projecta state logicalwindow as windowa state tablescanbuilder as tablescanbuilder projecta windowa windowa tablescanbuilder filter sql source account where age where age fields firstname age mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as state logicalfilter as state logicalrelation as relation projectb relation state after as after state logicalproject as projecta state tablescanbuilder as tablescanbuilder projecta tablescanbuilder exclamation tbd merge them remove duplicates or push down current behavior push down exclamation todo add its includehighlight optimize subqueries for sql and complex queries in ppl exclamation todo not implemented sample queries example from not optimized sql source account fields firstname lastname head mermaid statediagram state before as before state logicalproject as state logicallimit as limitb state logicalproject as state logicalrelation as relation limitb limitb relation state after as after state logicalproject as state logicallimit as limita state logicalproject as state tablescanbuilder as tablescanbuilder limita limita tablescanbuilder relevance search with subquery sql select highlight origin score from select origin as origin avgticketprice as price from flights where avgticketprice as f where score origin match query base this query fails because logicalfilter from outer query wasn’t pushed down so tried to do apply relevance search in memory unsupportedoperationexception opensearch defined function is only supported in where and having clause mermaid statediagram state before as before state logicalproject as state logicalhighlight as highlight state logicalfilter as state logicalproject as state logicalfilter as state logicalrelation as relation highlight highlight relation state after as after state logicalproject as state logicalfilter as filtera state logicalproject as state tablescanbuilder as tablescanbuilder filtera filtera tablescanbuilder complex ppl query returns incorrect results sql source account where age head sort balance where age fields age mermaid statediagram state before as before state logicalproject as projectb state logicalfilter as state logicalsort as state logicallimit as state logicalfilter as state logicalsort as state logicallimit as state logicalfilter as state logicalrelation as relation projectb relation state after as after state logicalproject as projecta state logicallimit as limita state tablescanbuilder as tablescanbuilder projecta limita limita tablescanbuilder optimizer rule update to satisfy requirement listed in new format rule format was created see example for push down filter below matches any amount of logical plan tree nodes of any types except logicalfilter mermaid statediagram state old push down filter implementation as oldfilter state logicalfilter as filterold state scanbuilder as relationold filterold relationold state new push down filter implementation as newfilter state logicalfilter as filternew state as dots state scanbuilder as relationnew filternew dots dots relationnew this new format was applied to push down page size push down filter push down aggregation push down filter push down sort push down highlight push down nested push down project and push down limit createtablescanbuilder createtablewritebuilder and prometheum related rules are not changed mermaid statediagram state createtablescanbuilder as builder state logicalrelation as scan pushdownrule class used to build push down rules the class architecture follows mermaid classdiagram class pushdownrule t class t clazz bifunction pushdownfunction list function exceptions boolean canbeappliedmultipletimes getdefaultexception function pushdownrule class t boolean bifunction pushdownrule class t boolean bifunction function pattern pattern t apply t captures logicalplan findtablescanbuilder logicalplan optional tablescanbuilder the following rule configurations are created rule tree node type can be applied multiple times push down function exception push down filter logicalfilter true pushdownfilter logicalaggregation logicalproject push down aggregation logicalaggregation false pushdownaggregation logicalproject push down sort logicalsort true pushdownsort logicalproject push down limit logicallimit false pushdownlimit logicalsort logicalfilter logicalproject push down project logicalproject false pushdownproject logicaleval logicalwindow push down highlight logicalhighlight true pushdownhighlight none push down nested logicalnested false pushdownnested none push down page size logicalpaginate false pushdownpagesize none note that default exception is applied to all rules rules are defined checked and applied in the following order push down page size push down filter push down aggregation push down filter push down sort push down highlight push down nested push down project push down limit to satisfy requirement push down filter is listed twice code please see code mvp poc in bit quill dev optimizer rework branch this code passes all it and contains only optimizer rework there re no new features so things marked as not implemented are still not implemented references current optimizer doc purpose proposed changes will unblock other fixes and features
0
746,505
26,032,684,886
IssuesEvent
2022-12-21 23:28:09
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
PreStart and PostStop event hooks
area/docker area/kubelet priority/awaiting-more-evidence sig/node kind/feature lifecycle/frozen
Many systems support event hooks for extensions. A few examples: https://developers.google.com/appengine/docs/java/javadoc/com/google/appengine/api/LifecycleManager http://developer.android.com/guide/components/activities.html http://upstart.ubuntu.com/cookbook/#event https://coreos.com/docs/launching-containers/launching/getting-started-with-systemd/ http://elasticbox.com/documentation/configuring-and-managing-boxes/start-stop-and-upgrade-boxes/ http://git-scm.com/docs/githooks.html docker stop and restart currently send SIGTERM followed by SIGKILL, similar to many other systems (e.g., Heroku: https://devcenter.heroku.com/articles/dynos#graceful-shutdown-with-sigterm), which provides an opportunity for applications to cleanly shut down, but lacks the ability to communicate the grace period duration or termination reason and doesn't directly provide support for notifying other processes or services. As described in the (liveness probe issue)[https://github.com/GoogleCloudPlatform/kubernetes/issues/66], it would be useful to support multiple types of hook execution/notification mechanisms. It would also be useful to pass arguments from clients, such as "reason" (e.g., "cancel", "restart", "reload", "resize", "reboot", "move", "host_update", "probe_failure"). Another way "reason" could be handled is with user-defined events. In addition to pre-termination notification, we should define other lifecycle hook points, probably at least pre- and post- start and terminate. It would be useful for post-terminate to be passed the (termination reason)[https://github.com/GoogleCloudPlatform/kubernetes/issues/137], which could either be successful completion, a client-provided stop reason (see above), or detailed failure reason (exit, signal, OOM, container creation error, docker crash, machine crash, lost/ghost). If the application generated an assertion failure or exception message, a post-termination hook could copy it to (/run/status.txt)[https://github.com/GoogleCloudPlatform/kubernetes/issues/139]. It would also be useful to be able to control (restart behavior)[https://github.com/GoogleCloudPlatform/kubernetes/issues/127] from a hook. We'd need a convenient way to carry over state from a previous execution. The simplest starting point would be for the user to keep it in a (volume)[https://github.com/GoogleCloudPlatform/kubernetes/issues/97].
1.0
PreStart and PostStop event hooks - Many systems support event hooks for extensions. A few examples: https://developers.google.com/appengine/docs/java/javadoc/com/google/appengine/api/LifecycleManager http://developer.android.com/guide/components/activities.html http://upstart.ubuntu.com/cookbook/#event https://coreos.com/docs/launching-containers/launching/getting-started-with-systemd/ http://elasticbox.com/documentation/configuring-and-managing-boxes/start-stop-and-upgrade-boxes/ http://git-scm.com/docs/githooks.html docker stop and restart currently send SIGTERM followed by SIGKILL, similar to many other systems (e.g., Heroku: https://devcenter.heroku.com/articles/dynos#graceful-shutdown-with-sigterm), which provides an opportunity for applications to cleanly shut down, but lacks the ability to communicate the grace period duration or termination reason and doesn't directly provide support for notifying other processes or services. As described in the (liveness probe issue)[https://github.com/GoogleCloudPlatform/kubernetes/issues/66], it would be useful to support multiple types of hook execution/notification mechanisms. It would also be useful to pass arguments from clients, such as "reason" (e.g., "cancel", "restart", "reload", "resize", "reboot", "move", "host_update", "probe_failure"). Another way "reason" could be handled is with user-defined events. In addition to pre-termination notification, we should define other lifecycle hook points, probably at least pre- and post- start and terminate. It would be useful for post-terminate to be passed the (termination reason)[https://github.com/GoogleCloudPlatform/kubernetes/issues/137], which could either be successful completion, a client-provided stop reason (see above), or detailed failure reason (exit, signal, OOM, container creation error, docker crash, machine crash, lost/ghost). If the application generated an assertion failure or exception message, a post-termination hook could copy it to (/run/status.txt)[https://github.com/GoogleCloudPlatform/kubernetes/issues/139]. It would also be useful to be able to control (restart behavior)[https://github.com/GoogleCloudPlatform/kubernetes/issues/127] from a hook. We'd need a convenient way to carry over state from a previous execution. The simplest starting point would be for the user to keep it in a (volume)[https://github.com/GoogleCloudPlatform/kubernetes/issues/97].
non_infrastructure
prestart and poststop event hooks many systems support event hooks for extensions a few examples docker stop and restart currently send sigterm followed by sigkill similar to many other systems e g heroku which provides an opportunity for applications to cleanly shut down but lacks the ability to communicate the grace period duration or termination reason and doesn t directly provide support for notifying other processes or services as described in the liveness probe issue it would be useful to support multiple types of hook execution notification mechanisms it would also be useful to pass arguments from clients such as reason e g cancel restart reload resize reboot move host update probe failure another way reason could be handled is with user defined events in addition to pre termination notification we should define other lifecycle hook points probably at least pre and post start and terminate it would be useful for post terminate to be passed the termination reason which could either be successful completion a client provided stop reason see above or detailed failure reason exit signal oom container creation error docker crash machine crash lost ghost if the application generated an assertion failure or exception message a post termination hook could copy it to run status txt it would also be useful to be able to control restart behavior from a hook we d need a convenient way to carry over state from a previous execution the simplest starting point would be for the user to keep it in a volume
0
35,355
31,034,111,399
IssuesEvent
2023-08-10 14:14:17
Star-ho/what-i-think
https://api.github.com/repos/Star-ho/what-i-think
opened
brotli
Infrastructure
google에서 개발한 압축방식 gzip 더 나은 압축성능을 보여줌 14% smaller JavaScript files 21% smaller HTML files 17% smaller CSS files IE는 지원안함...ㅠㅠ https://en.wikipedia.org/wiki/Brotli https://github.com/google/brotli https://www.siteground.com/blog/brotli-vs-gzip-compression/
1.0
brotli - google에서 개발한 압축방식 gzip 더 나은 압축성능을 보여줌 14% smaller JavaScript files 21% smaller HTML files 17% smaller CSS files IE는 지원안함...ㅠㅠ https://en.wikipedia.org/wiki/Brotli https://github.com/google/brotli https://www.siteground.com/blog/brotli-vs-gzip-compression/
infrastructure
brotli google에서 개발한 압축방식 gzip 더 나은 압축성능을 보여줌 smaller javascript files smaller html files smaller css files ie는 지원안함 ㅠㅠ
1
14,367
10,771,847,241
IssuesEvent
2019-11-02 10:44:39
ExchangeUnion/xud
https://api.github.com/repos/ExchangeUnion/xud
closed
Backup
P1 infrastructure
* Channel State backups: [SCB](https://github.com/lightningnetwork/lnd/blob/master/docs/recovery.md). LND fires an event whenever there is an update to the SCB file. Call: https://api.lightning.community/#subscribechannelbackups xud should then copy it to [location set by user](https://github.com/ExchangeUnion/xud/issues/246). Raiden: the channel state db should bin `~/.raiden` and be backed up the same way. Not sure if events are available. * `xud.db`, `xud.conf`, `lnd.conf`, `raiden conf`, `/root/.xud/tor` (xud container) Anything I forgot?
1.0
Backup - * Channel State backups: [SCB](https://github.com/lightningnetwork/lnd/blob/master/docs/recovery.md). LND fires an event whenever there is an update to the SCB file. Call: https://api.lightning.community/#subscribechannelbackups xud should then copy it to [location set by user](https://github.com/ExchangeUnion/xud/issues/246). Raiden: the channel state db should bin `~/.raiden` and be backed up the same way. Not sure if events are available. * `xud.db`, `xud.conf`, `lnd.conf`, `raiden conf`, `/root/.xud/tor` (xud container) Anything I forgot?
infrastructure
backup channel state backups lnd fires an event whenever there is an update to the scb file call xud should then copy it to raiden the channel state db should bin raiden and be backed up the same way not sure if events are available xud db xud conf lnd conf raiden conf root xud tor xud container anything i forgot
1
66,989
12,857,066,185
IssuesEvent
2020-07-09 08:42:17
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
opened
Update dependencies (1.45.0)
Code Quality change refine
<!-- If you've never submitted an issue to the SORMAS repository before or this is your first time using this template, please read the Contributing guidelines (accessible in the right sidebar) for an explanation about the information we'd like you to provide. --> ### Feature Description ### Problem Description Dependencies need updates at least to include bugfixes. ### Proposed Change 1. Hibernate 5 latest 2. Vaadin 8 latest 3. ... ### Possible Alternatives ### Additional Information
1.0
Update dependencies (1.45.0) - <!-- If you've never submitted an issue to the SORMAS repository before or this is your first time using this template, please read the Contributing guidelines (accessible in the right sidebar) for an explanation about the information we'd like you to provide. --> ### Feature Description ### Problem Description Dependencies need updates at least to include bugfixes. ### Proposed Change 1. Hibernate 5 latest 2. Vaadin 8 latest 3. ... ### Possible Alternatives ### Additional Information
non_infrastructure
update dependencies if you ve never submitted an issue to the sormas repository before or this is your first time using this template please read the contributing guidelines accessible in the right sidebar for an explanation about the information we d like you to provide feature description problem description dependencies need updates at least to include bugfixes proposed change hibernate latest vaadin latest possible alternatives additional information
0
314
2,642,716,383
IssuesEvent
2015-03-12 03:04:30
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
MSBuild consistency in casing boolean properties
:Infrastructure enhancement grabbed by assignee
MSBuild properties are case sensitive. The "false"/"true" vs "False"/"True" What do we prefer? It looks like we don't have any standards with this and I saw various places in the targets which mix two different casings even for the same properties. I believe we should unify this Examples: src\System.Collections.Immutable\src\System.Collections.Immutable.csproj (lowercase): ``` <PropertyGroup Condition=" '$(Configuration)|$(Platform)' == 'Debug|AnyCPU' "> <DebugSymbols>true</DebugSymbols> <DebugType>full</DebugType> <Optimize>false</Optimize> <DefineConstants>DEBUG;TRACE</DefineConstants> <ErrorReport>prompt</ErrorReport> <WarningLevel>4</WarningLevel> <Prefer32Bit>false</Prefer32Bit> <CodeAnalysisRuleSet>System.Collections.Immutable.ruleset</CodeAnalysisRuleSet> </PropertyGroup> ``` and the same project in the bottom: ``` <Compile Include="Strings.Designer.cs"> <AutoGen>True</AutoGen> <DesignTime>True</DesignTime> <DependentUpon>Strings.resx</DependentUpon> </Compile> ``` src\dirs.targets: (Look at the IsTestProject property and CLSCompliant condition): https://github.com/dotnet/corefx/blob/master/src/dir.targets#L94 ``` <PropertyGroup> <IsTestProject Condition="'$(IsTestProject)'=='' And $(MSBuildProjectName.EndsWith('.Tests'))">True</IsTestProject> <RunTestsForProject Condition="'$(RunTestsForProject)'=='' and '$(IsTestProject)'=='True'">True</RunTestsForProject> <RunTestsForProject Condition="'$(SkipTests)'=='True'">False</RunTestsForProject> <CLSCompliant Condition="'$(IsTestProject)'=='true'">false</CLSCompliant> </PropertyGroup> ```
1.0
MSBuild consistency in casing boolean properties - MSBuild properties are case sensitive. The "false"/"true" vs "False"/"True" What do we prefer? It looks like we don't have any standards with this and I saw various places in the targets which mix two different casings even for the same properties. I believe we should unify this Examples: src\System.Collections.Immutable\src\System.Collections.Immutable.csproj (lowercase): ``` <PropertyGroup Condition=" '$(Configuration)|$(Platform)' == 'Debug|AnyCPU' "> <DebugSymbols>true</DebugSymbols> <DebugType>full</DebugType> <Optimize>false</Optimize> <DefineConstants>DEBUG;TRACE</DefineConstants> <ErrorReport>prompt</ErrorReport> <WarningLevel>4</WarningLevel> <Prefer32Bit>false</Prefer32Bit> <CodeAnalysisRuleSet>System.Collections.Immutable.ruleset</CodeAnalysisRuleSet> </PropertyGroup> ``` and the same project in the bottom: ``` <Compile Include="Strings.Designer.cs"> <AutoGen>True</AutoGen> <DesignTime>True</DesignTime> <DependentUpon>Strings.resx</DependentUpon> </Compile> ``` src\dirs.targets: (Look at the IsTestProject property and CLSCompliant condition): https://github.com/dotnet/corefx/blob/master/src/dir.targets#L94 ``` <PropertyGroup> <IsTestProject Condition="'$(IsTestProject)'=='' And $(MSBuildProjectName.EndsWith('.Tests'))">True</IsTestProject> <RunTestsForProject Condition="'$(RunTestsForProject)'=='' and '$(IsTestProject)'=='True'">True</RunTestsForProject> <RunTestsForProject Condition="'$(SkipTests)'=='True'">False</RunTestsForProject> <CLSCompliant Condition="'$(IsTestProject)'=='true'">false</CLSCompliant> </PropertyGroup> ```
infrastructure
msbuild consistency in casing boolean properties msbuild properties are case sensitive the false true vs false true what do we prefer it looks like we don t have any standards with this and i saw various places in the targets which mix two different casings even for the same properties i believe we should unify this examples src system collections immutable src system collections immutable csproj lowercase true full false debug trace prompt false system collections immutable ruleset and the same project in the bottom true true strings resx src dirs targets look at the istestproject property and clscompliant condition true true false false
1
95,112
19,670,298,005
IssuesEvent
2022-01-11 06:11:18
J-J-B-J/LazySusan-Driver-Emulator
https://api.github.com/repos/J-J-B-J/LazySusan-Driver-Emulator
closed
User input validity checks can be replaced with try-except-else blocks
code improvement
This would make coding and reading code much easier.
1.0
User input validity checks can be replaced with try-except-else blocks - This would make coding and reading code much easier.
non_infrastructure
user input validity checks can be replaced with try except else blocks this would make coding and reading code much easier
0
13,072
10,112,918,839
IssuesEvent
2019-07-30 15:36:07
google/WebFundamentals
https://api.github.com/repos/google/WebFundamentals
closed
Bad anchor test doesn't detect bad anchors
type-Infrastructure
**Page Affected:** Any **What needs to be done?** There's an existing test for bad anchors, but it doesn't appear to catch all of them. The test is here: https://github.com/google/WebFundamentals/blob/master/gulp-tasks/tests/validateMarkdown.js#L281 If I understand correctly, \w will match alphanumeric characters and underscores, but not dashes. So it will catch this: `## Some heading {#some_heading}` But not this: `## Some heading {#some-heading}` - [X] Appropriate `type-something` label has been added. - [X] Appropriate `section-something` label has been added. (I don't think this section is specific to any section, so I didn't add one.)
1.0
Bad anchor test doesn't detect bad anchors - **Page Affected:** Any **What needs to be done?** There's an existing test for bad anchors, but it doesn't appear to catch all of them. The test is here: https://github.com/google/WebFundamentals/blob/master/gulp-tasks/tests/validateMarkdown.js#L281 If I understand correctly, \w will match alphanumeric characters and underscores, but not dashes. So it will catch this: `## Some heading {#some_heading}` But not this: `## Some heading {#some-heading}` - [X] Appropriate `type-something` label has been added. - [X] Appropriate `section-something` label has been added. (I don't think this section is specific to any section, so I didn't add one.)
infrastructure
bad anchor test doesn t detect bad anchors page affected any what needs to be done there s an existing test for bad anchors but it doesn t appear to catch all of them the test is here if i understand correctly w will match alphanumeric characters and underscores but not dashes so it will catch this some heading some heading but not this some heading some heading appropriate type something label has been added appropriate section something label has been added i don t think this section is specific to any section so i didn t add one
1
273,358
29,820,286,834
IssuesEvent
2023-06-17 01:21:38
pazhanivel07/frameworks_base_2021-0970
https://api.github.com/repos/pazhanivel07/frameworks_base_2021-0970
closed
CVE-2021-0439 (High) detected in baseandroid-10.0.0_r44 - autoclosed
Mend: dependency security vulnerability
## CVE-2021-0439 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/jni/com_android_server_power_PowerManagerService.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In setPowerModeWithHandle of com_android_server_power_PowerManagerService.cpp, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11Android ID: A-174243830 <p>Publish Date: 2021-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0439>CVE-2021-0439</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-04-01">https://source.android.com/security/bulletin/2021-04-01</a></p> <p>Release Date: 2022-02-13</p> <p>Fix Resolution: android-11.0.0_r34</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0439 (High) detected in baseandroid-10.0.0_r44 - autoclosed - ## CVE-2021-0439 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/jni/com_android_server_power_PowerManagerService.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In setPowerModeWithHandle of com_android_server_power_PowerManagerService.cpp, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11Android ID: A-174243830 <p>Publish Date: 2021-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0439>CVE-2021-0439</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-04-01">https://source.android.com/security/bulletin/2021-04-01</a></p> <p>Release Date: 2022-02-13</p> <p>Fix Resolution: android-11.0.0_r34</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in baseandroid autoclosed cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in base branch master vulnerable source files services core jni com android server power powermanagerservice cpp vulnerability details in setpowermodewithhandle of com android server power powermanagerservice cpp there is a possible out of bounds write due to a missing bounds check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
31,571
25,891,707,350
IssuesEvent
2022-12-14 18:30:47
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
outerloop legs failing to detect `dotnet`
area-Infrastructure-libraries blocking-outerloop in-pr
Original issue was opened in dotnet/arcade: https://github.com/dotnet/arcade/issues/11185 Many outerloop legs fail with this message when attempting to run unit tests: ``` ./RunTests.sh: line 168: /tmp/helix/working/A6B808E3/p/dotnet: No such file or directory ``` The investigation found that this may have been caused by a runtime change that happened between these two runs (from 8/2 to 8/3): pr/public/dotnet/runtime/refs/pull/73300/merge pr/public/dotnet/runtime/refs/pull/73020/merge <details> <summary>Kusto Query</summary> ```kusto Jobs | where QueueName == "osx.1200.amd64.open" | extend Propz = parse_json(Properties) | where Propz["DefinitionName"] == "runtime-libraries-coreclr outerloop" | where Propz["System.PhaseName"] == "libraries_build_OSX_x64_Debug" | order by Finished desc ``` </details> The main suspect PR is one that seeked to remove the host packages conditionally and was merged at a time between the above two PRs likely merge commits: https://github.com/dotnet/runtime/pull/73095/files We need to confirm this issue caused the regression and get it fixed.
1.0
outerloop legs failing to detect `dotnet` - Original issue was opened in dotnet/arcade: https://github.com/dotnet/arcade/issues/11185 Many outerloop legs fail with this message when attempting to run unit tests: ``` ./RunTests.sh: line 168: /tmp/helix/working/A6B808E3/p/dotnet: No such file or directory ``` The investigation found that this may have been caused by a runtime change that happened between these two runs (from 8/2 to 8/3): pr/public/dotnet/runtime/refs/pull/73300/merge pr/public/dotnet/runtime/refs/pull/73020/merge <details> <summary>Kusto Query</summary> ```kusto Jobs | where QueueName == "osx.1200.amd64.open" | extend Propz = parse_json(Properties) | where Propz["DefinitionName"] == "runtime-libraries-coreclr outerloop" | where Propz["System.PhaseName"] == "libraries_build_OSX_x64_Debug" | order by Finished desc ``` </details> The main suspect PR is one that seeked to remove the host packages conditionally and was merged at a time between the above two PRs likely merge commits: https://github.com/dotnet/runtime/pull/73095/files We need to confirm this issue caused the regression and get it fixed.
infrastructure
outerloop legs failing to detect dotnet original issue was opened in dotnet arcade many outerloop legs fail with this message when attempting to run unit tests runtests sh line tmp helix working p dotnet no such file or directory the investigation found that this may have been caused by a runtime change that happened between these two runs from to pr public dotnet runtime refs pull merge pr public dotnet runtime refs pull merge kusto query kusto jobs where queuename osx open extend propz parse json properties where propz runtime libraries coreclr outerloop where propz libraries build osx debug order by finished desc the main suspect pr is one that seeked to remove the host packages conditionally and was merged at a time between the above two prs likely merge commits we need to confirm this issue caused the regression and get it fixed
1
63,384
8,677,777,386
IssuesEvent
2018-11-30 17:46:49
prkumar/uplink
https://api.github.com/repos/prkumar/uplink
closed
Small issue in docs
Documentation good first issue help wanted
I think there's a minor issue with this snippet in the docs at https://uplink.readthedocs.io/en/stable/user/quickstart.html#response-and-error-handling ``` def raise_for_status(response): """Checks whether or not the response was successful.""" if 200 <= response.status <= 299: raise UnsuccessfulRequest(response.url) # Pass through the response. return response ``` I think condition of `if` describes a success case rather than a failure. Also if response is `response` of requests library it should probably be `response.status_code` (Not sure whether `response.status` also works or not. I don't think it does.)
1.0
Small issue in docs - I think there's a minor issue with this snippet in the docs at https://uplink.readthedocs.io/en/stable/user/quickstart.html#response-and-error-handling ``` def raise_for_status(response): """Checks whether or not the response was successful.""" if 200 <= response.status <= 299: raise UnsuccessfulRequest(response.url) # Pass through the response. return response ``` I think condition of `if` describes a success case rather than a failure. Also if response is `response` of requests library it should probably be `response.status_code` (Not sure whether `response.status` also works or not. I don't think it does.)
non_infrastructure
small issue in docs i think there s a minor issue with this snippet in the docs at def raise for status response checks whether or not the response was successful if response status raise unsuccessfulrequest response url pass through the response return response i think condition of if describes a success case rather than a failure also if response is response of requests library it should probably be response status code not sure whether response status also works or not i don t think it does
0
189,682
14,517,656,373
IssuesEvent
2020-12-13 20:28:45
opencv/opencv
https://api.github.com/repos/opencv/opencv
opened
Nightly: test_gapi_infer.py failed on Skylake-X platform (AVX512)
bug category: dnn category: g-api / gapi category: python bindings test
[Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-skx-lin64/builds/869). ``` ====================================================================== FAIL: test_age_gender_infer (test_gapi_infer.test_gapi_infer) ---------------------------------------------------------------------- Traceback (most recent call last): File "/build/master_openvino-skx-lin64/opencv/modules/gapi/misc/python/test/test_gapi_infer.py", line 57, in test_age_gender_infer self.assertEqual(0.0, cv.norm(dnn_gender, gapi_gender, cv.NORM_INF)) AssertionError: 0.0 != 2.7939677238464355e-08 ```
1.0
Nightly: test_gapi_infer.py failed on Skylake-X platform (AVX512) - [Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-skx-lin64/builds/869). ``` ====================================================================== FAIL: test_age_gender_infer (test_gapi_infer.test_gapi_infer) ---------------------------------------------------------------------- Traceback (most recent call last): File "/build/master_openvino-skx-lin64/opencv/modules/gapi/misc/python/test/test_gapi_infer.py", line 57, in test_age_gender_infer self.assertEqual(0.0, cv.norm(dnn_gender, gapi_gender, cv.NORM_INF)) AssertionError: 0.0 != 2.7939677238464355e-08 ```
non_infrastructure
nightly test gapi infer py failed on skylake x platform fail test age gender infer test gapi infer test gapi infer traceback most recent call last file build master openvino skx opencv modules gapi misc python test test gapi infer py line in test age gender infer self assertequal cv norm dnn gender gapi gender cv norm inf assertionerror
0
618
2,794,096,904
IssuesEvent
2015-05-11 14:59:17
insieme/insieme
https://api.github.com/repos/insieme/insieme
closed
Cleanup build process by removing own GCC dependencies
infrastructure setup issue
I'm not completely sure, but I think we finally got to the point where we can abandon our GCC and ruby builds in the lib directory (or make it a opt-in option) and use the system compiler. The C++11 standard should be widely enough spread already, now that actually its successor is already out. The reason for this: our README front page for the project is very complex, in particular the part with the LD_LIBRARY_PATHS is anoyingly complex; this could help to eliminate this obstacle for future users.
1.0
Cleanup build process by removing own GCC dependencies - I'm not completely sure, but I think we finally got to the point where we can abandon our GCC and ruby builds in the lib directory (or make it a opt-in option) and use the system compiler. The C++11 standard should be widely enough spread already, now that actually its successor is already out. The reason for this: our README front page for the project is very complex, in particular the part with the LD_LIBRARY_PATHS is anoyingly complex; this could help to eliminate this obstacle for future users.
infrastructure
cleanup build process by removing own gcc dependencies i m not completely sure but i think we finally got to the point where we can abandon our gcc and ruby builds in the lib directory or make it a opt in option and use the system compiler the c standard should be widely enough spread already now that actually its successor is already out the reason for this our readme front page for the project is very complex in particular the part with the ld library paths is anoyingly complex this could help to eliminate this obstacle for future users
1