Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
325,520
27,883,540,355
IssuesEvent
2023-03-21 21:25:33
primitivefinance/arbiter
https://api.github.com/repos/primitivefinance/arbiter
closed
Let agents get events from contract interactions
Agents Testing & Error Handling
We want agents to be able to pick up on events that contracts emit, e.g., a change in price on a DEX. - [ ] Create a test that displays this functionality.
1.0
Let agents get events from contract interactions - We want agents to be able to pick up on events that contracts emit, e.g., a change in price on a DEX. - [ ] Create a test that displays this functionality.
test
let agents get events from contract interactions we want agents to be able to pick up on events that contracts emit e g a change in price on a dex create a test that displays this functionality
1
153,723
12,158,629,139
IssuesEvent
2020-04-26 05:07:19
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Jest Integration Tests.src/core/server/legacy/integration_tests - logging service compatibility logging config respects legacy logging settings "silent": true
failed-test
A test failed on a tracked branch ``` Error: Http server is not setup up yet at HttpServer.start (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/src/core/server/http/http_server.ts:145:13) at HttpService.start (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/src/core/server/http/http_service.ts:134:29) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/4693/) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/legacy/integration_tests","test.name":"logging service compatibility logging config respects legacy logging settings \"silent\": true","test.failCount":19}} -->
1.0
Failing test: Jest Integration Tests.src/core/server/legacy/integration_tests - logging service compatibility logging config respects legacy logging settings "silent": true - A test failed on a tracked branch ``` Error: Http server is not setup up yet at HttpServer.start (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/src/core/server/http/http_server.ts:145:13) at HttpService.start (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/src/core/server/http/http_service.ts:134:29) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/4693/) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/legacy/integration_tests","test.name":"logging service compatibility logging config respects legacy logging settings \"silent\": true","test.failCount":19}} -->
test
failing test jest integration tests src core server legacy integration tests logging service compatibility logging config respects legacy logging settings silent true a test failed on a tracked branch error http server is not setup up yet at httpserver start var lib jenkins workspace elastic kibana master kibana src core server http http server ts at httpservice start var lib jenkins workspace elastic kibana master kibana src core server http http service ts first failure
1
814
10,632,192,697
IssuesEvent
2019-10-15 09:51:47
Cha-OS/colabo
https://api.github.com/repos/Cha-OS/colabo
opened
resources-to-be-loaded for "User Actions Statuses"
moderation reliability visual_improvement
to provide "loading 1/5 ..." text and to hide refreshing/loading info only when all resources are loaded to provide the method for informing when all resources are loaded
True
resources-to-be-loaded for "User Actions Statuses" - to provide "loading 1/5 ..." text and to hide refreshing/loading info only when all resources are loaded to provide the method for informing when all resources are loaded
non_test
resources to be loaded for user actions statuses to provide loading text and to hide refreshing loading info only when all resources are loaded to provide the method for informing when all resources are loaded
0
25,996
4,188,116,399
IssuesEvent
2016-06-23 19:39:33
Microsoft/RTVS
https://api.github.com/repos/Microsoft/RTVS
reopened
Host related tests fail in batch with timeout
type:bug type:test issue
That is caused by C0000135 exit in the host (c0000135 is "Unable To Locate Component" ) ![image](https://cloud.githubusercontent.com/assets/12820357/16287985/2f52f530-389c-11e6-96ec-c68caa3312ac.png)
1.0
Host related tests fail in batch with timeout - That is caused by C0000135 exit in the host (c0000135 is "Unable To Locate Component" ) ![image](https://cloud.githubusercontent.com/assets/12820357/16287985/2f52f530-389c-11e6-96ec-c68caa3312ac.png)
test
host related tests fail in batch with timeout that is caused by exit in the host is unable to locate component
1
6,243
6,266,264,709
IssuesEvent
2017-07-17 00:35:29
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Let's Encrypt integration does not support specific IP bindings
area/infrastructure bug
Source: http://help.octopusdeploy.com/discussions/problems/55745 Currently the Let's Encrypt integration is configured to bind certificates on 0.0.0.0, which is for any IP bound to the server. This can be a problem when trying to run additional Octopus Server instances (or IIS) on the same server. Provide the ability to specify IP address when configuring Let's Encrypt.
1.0
Let's Encrypt integration does not support specific IP bindings - Source: http://help.octopusdeploy.com/discussions/problems/55745 Currently the Let's Encrypt integration is configured to bind certificates on 0.0.0.0, which is for any IP bound to the server. This can be a problem when trying to run additional Octopus Server instances (or IIS) on the same server. Provide the ability to specify IP address when configuring Let's Encrypt.
non_test
let s encrypt integration does not support specific ip bindings source currently the let s encrypt integration is configured to bind certificates on which is for any ip bound to the server this can be a problem when trying to run additional octopus server instances or iis on the same server provide the ability to specify ip address when configuring let s encrypt
0
280,062
24,278,002,890
IssuesEvent
2022-09-28 15:07:58
PublicDataWorks/complaint-manager
https://api.github.com/repos/PublicDataWorks/complaint-manager
closed
Running lighthouse against CI
improve testing 🧪
**Describe the bug** Lighthouse is running against demo (urls are found in `lighthouserc.json`). We want to run it against CI. The issue we face is we get stuck on the login page when using CI urls. We want to bypass this. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'lighthousesrc.json' 2. Replace the demo URLS to point against CI. 3. Push the code. 4. Go to the lighthouse job in pipeline 5. Expand the "lhci autorun" tab 6. View the accessibility error and find the corresponding link to each warning 7. You'll see a small screenshot of login page instead of the actual page desired for accessibility checks. **Expected behavior** 1. To be able to mimic the behavior when lighthouse is ran against demo. (Having no issues with authentication) **Additional context** https://github.com/GoogleChrome/lighthouse-ci/blob/main/docs/configuration.md We can try using puppeteerScript. Description and steps are in the link above.
1.0
Running lighthouse against CI - **Describe the bug** Lighthouse is running against demo (urls are found in `lighthouserc.json`). We want to run it against CI. The issue we face is we get stuck on the login page when using CI urls. We want to bypass this. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'lighthousesrc.json' 2. Replace the demo URLS to point against CI. 3. Push the code. 4. Go to the lighthouse job in pipeline 5. Expand the "lhci autorun" tab 6. View the accessibility error and find the corresponding link to each warning 7. You'll see a small screenshot of login page instead of the actual page desired for accessibility checks. **Expected behavior** 1. To be able to mimic the behavior when lighthouse is ran against demo. (Having no issues with authentication) **Additional context** https://github.com/GoogleChrome/lighthouse-ci/blob/main/docs/configuration.md We can try using puppeteerScript. Description and steps are in the link above.
test
running lighthouse against ci describe the bug lighthouse is running against demo urls are found in lighthouserc json we want to run it against ci the issue we face is we get stuck on the login page when using ci urls we want to bypass this to reproduce steps to reproduce the behavior go to lighthousesrc json replace the demo urls to point against ci push the code go to the lighthouse job in pipeline expand the lhci autorun tab view the accessibility error and find the corresponding link to each warning you ll see a small screenshot of login page instead of the actual page desired for accessibility checks expected behavior to be able to mimic the behavior when lighthouse is ran against demo having no issues with authentication additional context we can try using puppeteerscript description and steps are in the link above
1
224,519
17,753,739,988
IssuesEvent
2021-08-28 10:19:47
sot/proseco
https://api.github.com/repos/sot/proseco
closed
API: provide obsid to proseco
enhancement Matlab testing
In the context of catalog checking (thumbs up) there can be different rules for ORs and ERs. An obsid is the normal way of differentiating. Strictly speaking this might be needed only for aca_review, but putting it upstream in the proseco ACA table object would be nice.
1.0
API: provide obsid to proseco - In the context of catalog checking (thumbs up) there can be different rules for ORs and ERs. An obsid is the normal way of differentiating. Strictly speaking this might be needed only for aca_review, but putting it upstream in the proseco ACA table object would be nice.
test
api provide obsid to proseco in the context of catalog checking thumbs up there can be different rules for ors and ers an obsid is the normal way of differentiating strictly speaking this might be needed only for aca review but putting it upstream in the proseco aca table object would be nice
1
137,536
11,140,404,598
IssuesEvent
2019-12-21 14:08:27
dexpenses/dexpenses-extract
https://api.github.com/repos/dexpenses/dexpenses-extract
closed
Implement test receipt normal/hannover-dinea-cash
enhancement test-data
Receipt to implement: ![normal/hannover-dinea-cash](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/normal%2Fhannover-dinea-cash.JPEG?alt=media "normal/hannover-dinea-cash")
1.0
Implement test receipt normal/hannover-dinea-cash - Receipt to implement: ![normal/hannover-dinea-cash](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/normal%2Fhannover-dinea-cash.JPEG?alt=media "normal/hannover-dinea-cash")
test
implement test receipt normal hannover dinea cash receipt to implement normal hannover dinea cash
1
66,617
27,528,009,882
IssuesEvent
2023-03-06 19:37:45
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Missing Migration steps from "old" aad-pod-identity
container-service/svc triaged assigned-to-author doc-enhancement Pri2
since this is preview feature , most of the users of AAD-Pod-Identity come from https://azure.github.io/aad-pod-identity/docs/demo/standard_walkthrough/ thus migration steps from that "manual" mode to this new "managed" mode is much required. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: c3c187c0-27f1-11c9-41af-6c65ca36e77e * Version Independent ID: 249a570b-90ba-4a11-4467-320f0369ec25 * Content: [Use Azure Active Directory pod-managed identities in Azure Kubernetes Service (Preview) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/use-azure-ad-pod-identity) * Content Source: [articles/aks/use-azure-ad-pod-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/use-azure-ad-pod-identity.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
1.0
Missing Migration steps from "old" aad-pod-identity - since this is preview feature , most of the users of AAD-Pod-Identity come from https://azure.github.io/aad-pod-identity/docs/demo/standard_walkthrough/ thus migration steps from that "manual" mode to this new "managed" mode is much required. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: c3c187c0-27f1-11c9-41af-6c65ca36e77e * Version Independent ID: 249a570b-90ba-4a11-4467-320f0369ec25 * Content: [Use Azure Active Directory pod-managed identities in Azure Kubernetes Service (Preview) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/use-azure-ad-pod-identity) * Content Source: [articles/aks/use-azure-ad-pod-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/use-azure-ad-pod-identity.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
non_test
missing migration steps from old aad pod identity since this is preview feature most of the users of aad pod identity come from thus migration steps from that manual mode to this new managed mode is much required document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login zr msft microsoft alias zarhoads
0
50,576
6,402,316,438
IssuesEvent
2017-08-06 08:08:04
aeonlang/designs
https://api.github.com/repos/aeonlang/designs
opened
File Structures
Compiler Design
## Intro This document describes the initial document structures for projects and solutions. ## Pretext The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this document are to be interpreted as described in [RFC 2119](https://www.ietf.org/rfc/rfc2119.txt). ## aeon.json `aeon.json` is the solution level file that has the following format: { "projectDirectories": ["src", "tests"], "targetFrameworks": ["win_x64", "win_x86", "MSIL"], "buildConfigurations": [ // anything set in here gets inherited by all projects in the src/tests directory { "name": "debug", "compilationOptions": { "applyOptimizations": false, "compiledFiles": ["**/*.fli"], // default "excludedFiles": ["**/*.dll"], // for example, default is empty "copiedFiles": ["**/*.png"] // files to be copied to output directory, default is empty } } ] } ## aeonproj.json `aeonproj.json` is the project level file that has the following format: { "packageName": "fli.sample", "version": "1.0.0", // semver "authors": [{ "name": "Stuart Blackler", // required "email": "not@areal.email", // optional "website": "https://im5tu.io" // optional }], "licenseUrl": "https://im5tu.io", // required "projectUrl": "https://im5tu.io", // required } ## aeontool.json `aeontool.json` is the tool definition file that has the following format: { toolName: "aeon-compile", repository: "git@github.com:aeonlang/compiler.git" configuration: [ { key: "", value: "" } ] } Properties: - `toolName`: The name of the executable (minus extensions) that is ran. - `configuration`: Default key value pairs which are passed to the application when called from `aeon.exe`
1.0
File Structures - ## Intro This document describes the initial document structures for projects and solutions. ## Pretext The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this document are to be interpreted as described in [RFC 2119](https://www.ietf.org/rfc/rfc2119.txt). ## aeon.json `aeon.json` is the solution level file that has the following format: { "projectDirectories": ["src", "tests"], "targetFrameworks": ["win_x64", "win_x86", "MSIL"], "buildConfigurations": [ // anything set in here gets inherited by all projects in the src/tests directory { "name": "debug", "compilationOptions": { "applyOptimizations": false, "compiledFiles": ["**/*.fli"], // default "excludedFiles": ["**/*.dll"], // for example, default is empty "copiedFiles": ["**/*.png"] // files to be copied to output directory, default is empty } } ] } ## aeonproj.json `aeonproj.json` is the project level file that has the following format: { "packageName": "fli.sample", "version": "1.0.0", // semver "authors": [{ "name": "Stuart Blackler", // required "email": "not@areal.email", // optional "website": "https://im5tu.io" // optional }], "licenseUrl": "https://im5tu.io", // required "projectUrl": "https://im5tu.io", // required } ## aeontool.json `aeontool.json` is the tool definition file that has the following format: { toolName: "aeon-compile", repository: "git@github.com:aeonlang/compiler.git" configuration: [ { key: "", value: "" } ] } Properties: - `toolName`: The name of the executable (minus extensions) that is ran. - `configuration`: Default key value pairs which are passed to the application when called from `aeon.exe`
non_test
file structures intro this document describes the initial document structures for projects and solutions pretext the key words must must not required shall shall should should not recommended may and optional in this document are to be interpreted as described in aeon json aeon json is the solution level file that has the following format projectdirectories targetframeworks buildconfigurations anything set in here gets inherited by all projects in the src tests directory name debug compilationoptions applyoptimizations false compiledfiles default excludedfiles for example default is empty copiedfiles files to be copied to output directory default is empty aeonproj json aeonproj json is the project level file that has the following format packagename fli sample version semver authors name stuart blackler required email not areal email optional website optional licenseurl required projecturl required aeontool json aeontool json is the tool definition file that has the following format toolname aeon compile repository git github com aeonlang compiler git configuration key value properties toolname the name of the executable minus extensions that is ran configuration default key value pairs which are passed to the application when called from aeon exe
0
123,162
10,255,965,593
IssuesEvent
2019-08-21 16:31:29
spack/spack
https://api.github.com/repos/spack/spack
closed
`spack pkg` command and tests fail for git < 1.8.5
bug robustness tests
The implementation of the `spack pkg` command assumes one is working in an environment with a git >= 1.8.5, as the `-C` option is used for all invocations of the `git` command. Associated tests also fail, see `spack test cmd/pkg`: ```console $ spack test cmd/pkg ... E ProcessError: Command exited with status 129: E '/usr/bin/git' '-C' '/tmp/pytest-of-greenc/pytest-52/mock_pkg_git_repo0/builtin.mock/packages' 'ls-tree' '--name-only' 'HEAD^^' ../util/executable.py:189: ProcessError -------------------------------------------------------------------------- Captured stderr call --------------------------------------------------------------------------- Unknown option: -C usage: git [--version] [--help] [-c name=value] [--exec-path[=<path>]] [--html-path] [--man-path] [--info-path] [-p|--paginate|--no-pager] [--no-replace-objects] [--bare] [--git-dir=<path>] [--work-tree=<path>] [--namespace=<name>] <command> [<args>] ``` On a standard `Scientific7` (or `RHEL7`) system, the version of git available is `1.8.3.1`. On `RHEL6` (still in extensive use in HEP), it is `1.7.1`. A reasonable solution might be to test for the version and fall back on a function wrapper to the git `Executable` that invokes it with the current directory corresponding to the desired `path`.
1.0
`spack pkg` command and tests fail for git < 1.8.5 - The implementation of the `spack pkg` command assumes one is working in an environment with a git >= 1.8.5, as the `-C` option is used for all invocations of the `git` command. Associated tests also fail, see `spack test cmd/pkg`: ```console $ spack test cmd/pkg ... E ProcessError: Command exited with status 129: E '/usr/bin/git' '-C' '/tmp/pytest-of-greenc/pytest-52/mock_pkg_git_repo0/builtin.mock/packages' 'ls-tree' '--name-only' 'HEAD^^' ../util/executable.py:189: ProcessError -------------------------------------------------------------------------- Captured stderr call --------------------------------------------------------------------------- Unknown option: -C usage: git [--version] [--help] [-c name=value] [--exec-path[=<path>]] [--html-path] [--man-path] [--info-path] [-p|--paginate|--no-pager] [--no-replace-objects] [--bare] [--git-dir=<path>] [--work-tree=<path>] [--namespace=<name>] <command> [<args>] ``` On a standard `Scientific7` (or `RHEL7`) system, the version of git available is `1.8.3.1`. On `RHEL6` (still in extensive use in HEP), it is `1.7.1`. A reasonable solution might be to test for the version and fall back on a function wrapper to the git `Executable` that invokes it with the current directory corresponding to the desired `path`.
test
spack pkg command and tests fail for git as the c option is used for all invocations of the git command associated tests also fail see spack test cmd pkg console spack test cmd pkg e processerror command exited with status e usr bin git c tmp pytest of greenc pytest mock pkg git builtin mock packages ls tree name only head util executable py processerror captured stderr call unknown option c usage git on a standard or system the version of git available is on still in extensive use in hep it is a reasonable solution might be to test for the version and fall back on a function wrapper to the git executable that invokes it with the current directory corresponding to the desired path
1
89,236
8,198,466,482
IssuesEvent
2018-08-31 16:29:31
aspnet/Hosting
https://api.github.com/repos/aspnet/Hosting
closed
Test failure: ShutdownTestRun
Branch:2.2 Branch:master test-failure
This test [fails](http://aspnetci/viewLog.html?buildId=528440&buildTypeId=Releases_22xPublic_UbuntuUniverse) occasionally with the following error: ``` System.NullReferenceException : Object reference not set to an instance of an object. at Microsoft.AspNetCore.Server.IntegrationTesting.ApplicationDeployer.CleanPublishedOutput() in /_/src/Microsoft.AspNetCore.Server.IntegrationTesting/Deployers/ApplicationDeployer.cs:line 100 at Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer.Dispose() in /_/src/Microsoft.AspNetCore.Server.IntegrationTesting/Deployers/SelfHostDeployer.cs:line 205 at Microsoft.AspNetCore.Hosting.FunctionalTests.ShutdownTests.ExecuteShutdownTest(String testName, String shutdownMechanic) in /_/test/Microsoft.AspNetCore.Hosting.FunctionalTests/ShutdownTests.cs:line 114 at Microsoft.AspNetCore.Hosting.FunctionalTests.ShutdownTests.ShutdownTestRun() in /_/test/Microsoft.AspNetCore.Hosting.FunctionalTests/ShutdownTests.cs:line 33 --- End of stack trace from previous location where exception was thrown --- ------- Stdout: ------- | [0.015s] TestLifetime Information: Starting test ShutdownTestRun at 2018-08-10T22:40:24 | [0.001s] TestLifetime Information: Starting test ExecuteShutdownTest at 2018-08-10T22:40:24 | [0.009s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: Deploying [Variation] :: ServerType=Kestrel, Runtime=CoreClr, Arch=x64, BaseUrlHint=, Publish=True | [0.014s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: Executing command dotnet publish --output "/mnt/temp/buildTmp/27932f56ea2345cc9813f135d51da328" --framework netcoreapp2.2 --configuration Release --no-restore -p:VerifyMatchingImplicitPackageVersion=false | [0.332s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: Microsoft (R) Build Engine version 15.7.179.6572 for .NET Core | [0.333s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: Copyright (C) Microsoft Corporation. All rights reserved. | [2.505s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/GenerateAssemblyInfo.targets(25,5): error MSB4062: The "Microsoft.AspNetCore.BuildTools.Sdk_GetAssemblyFileVersion" task could not be loaded from the assembly /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/../tools/netcoreapp2.1/Internal.AspNetCore.Sdk.dll. Assembly with same name is already loaded Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask. [/mnt/work/33bdfc1cae7b2a38/modules/Hosting/src/Microsoft.Extensions.Hosting.Abstractions/Microsoft.Extensions.Hosting.Abstractions.csproj] | [2.515s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/GenerateAssemblyInfo.targets(25,5): error MSB4062: The "Microsoft.AspNetCore.BuildTools.Sdk_GetAssemblyFileVersion" task could not be loaded from the assembly /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/../tools/netcoreapp2.1/Internal.AspNetCore.Sdk.dll. Assembly with same name is already loaded Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask. [/mnt/work/33bdfc1cae7b2a38/modules/Hosting/src/Microsoft.AspNetCore.Hosting.Server.Abstractions/Microsoft.AspNetCore.Hosting.Server.Abstractions.csproj] | [2.602s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Error: dotnet publish exited with exit code : 1 | [2.605s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Warning: Host process already exited or never started successfully. | [2.607s] TestLifetime Information: Finished test ExecuteShutdownTest in 2.6064648s | [2.664s] TestLifetime Information: Finished test ShutdownTestRun in 2.6582765s ``` Other tests within that build may have failed with a similar message, but they are not listed here. Check the link above for more info. This test failed on 2.2. CC @Eilon (because the bot doesn't know who else to pick) This issue was made automatically. If there is a problem contact ryanbrandenburg.
1.0
Test failure: ShutdownTestRun - This test [fails](http://aspnetci/viewLog.html?buildId=528440&buildTypeId=Releases_22xPublic_UbuntuUniverse) occasionally with the following error: ``` System.NullReferenceException : Object reference not set to an instance of an object. at Microsoft.AspNetCore.Server.IntegrationTesting.ApplicationDeployer.CleanPublishedOutput() in /_/src/Microsoft.AspNetCore.Server.IntegrationTesting/Deployers/ApplicationDeployer.cs:line 100 at Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer.Dispose() in /_/src/Microsoft.AspNetCore.Server.IntegrationTesting/Deployers/SelfHostDeployer.cs:line 205 at Microsoft.AspNetCore.Hosting.FunctionalTests.ShutdownTests.ExecuteShutdownTest(String testName, String shutdownMechanic) in /_/test/Microsoft.AspNetCore.Hosting.FunctionalTests/ShutdownTests.cs:line 114 at Microsoft.AspNetCore.Hosting.FunctionalTests.ShutdownTests.ShutdownTestRun() in /_/test/Microsoft.AspNetCore.Hosting.FunctionalTests/ShutdownTests.cs:line 33 --- End of stack trace from previous location where exception was thrown --- ------- Stdout: ------- | [0.015s] TestLifetime Information: Starting test ShutdownTestRun at 2018-08-10T22:40:24 | [0.001s] TestLifetime Information: Starting test ExecuteShutdownTest at 2018-08-10T22:40:24 | [0.009s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: Deploying [Variation] :: ServerType=Kestrel, Runtime=CoreClr, Arch=x64, BaseUrlHint=, Publish=True | [0.014s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: Executing command dotnet publish --output "/mnt/temp/buildTmp/27932f56ea2345cc9813f135d51da328" --framework netcoreapp2.2 --configuration Release --no-restore -p:VerifyMatchingImplicitPackageVersion=false | [0.332s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: Microsoft (R) Build Engine version 15.7.179.6572 for .NET Core | [0.333s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: Copyright (C) Microsoft Corporation. All rights reserved. | [2.505s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/GenerateAssemblyInfo.targets(25,5): error MSB4062: The "Microsoft.AspNetCore.BuildTools.Sdk_GetAssemblyFileVersion" task could not be loaded from the assembly /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/../tools/netcoreapp2.1/Internal.AspNetCore.Sdk.dll. Assembly with same name is already loaded Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask. [/mnt/work/33bdfc1cae7b2a38/modules/Hosting/src/Microsoft.Extensions.Hosting.Abstractions/Microsoft.Extensions.Hosting.Abstractions.csproj] | [2.515s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Information: dotnet-publish stdout: /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/GenerateAssemblyInfo.targets(25,5): error MSB4062: The "Microsoft.AspNetCore.BuildTools.Sdk_GetAssemblyFileVersion" task could not be loaded from the assembly /mnt/work/33bdfc1cae7b2a38/.packages/internal.aspnetcore.sdk/2.2.0-preview1-20180807.2/build/../tools/netcoreapp2.1/Internal.AspNetCore.Sdk.dll. Assembly with same name is already loaded Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask. [/mnt/work/33bdfc1cae7b2a38/modules/Hosting/src/Microsoft.AspNetCore.Hosting.Server.Abstractions/Microsoft.AspNetCore.Hosting.Server.Abstractions.csproj] | [2.602s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Error: dotnet publish exited with exit code : 1 | [2.605s] Microsoft.AspNetCore.Server.IntegrationTesting.SelfHostDeployer Warning: Host process already exited or never started successfully. | [2.607s] TestLifetime Information: Finished test ExecuteShutdownTest in 2.6064648s | [2.664s] TestLifetime Information: Finished test ShutdownTestRun in 2.6582765s ``` Other tests within that build may have failed with a similar message, but they are not listed here. Check the link above for more info. This test failed on 2.2. CC @Eilon (because the bot doesn't know who else to pick) This issue was made automatically. If there is a problem contact ryanbrandenburg.
test
test failure shutdowntestrun this test occasionally with the following error system nullreferenceexception object reference not set to an instance of an object at microsoft aspnetcore server integrationtesting applicationdeployer cleanpublishedoutput in src microsoft aspnetcore server integrationtesting deployers applicationdeployer cs line at microsoft aspnetcore server integrationtesting selfhostdeployer dispose in src microsoft aspnetcore server integrationtesting deployers selfhostdeployer cs line at microsoft aspnetcore hosting functionaltests shutdowntests executeshutdowntest string testname string shutdownmechanic in test microsoft aspnetcore hosting functionaltests shutdowntests cs line at microsoft aspnetcore hosting functionaltests shutdowntests shutdowntestrun in test microsoft aspnetcore hosting functionaltests shutdowntests cs line end of stack trace from previous location where exception was thrown stdout testlifetime information starting test shutdowntestrun at testlifetime information starting test executeshutdowntest at microsoft aspnetcore server integrationtesting selfhostdeployer information deploying servertype kestrel runtime coreclr arch baseurlhint publish true microsoft aspnetcore server integrationtesting selfhostdeployer information executing command dotnet publish output mnt temp buildtmp framework configuration release no restore p verifymatchingimplicitpackageversion false microsoft aspnetcore server integrationtesting selfhostdeployer information dotnet publish stdout microsoft r build engine version for net core microsoft aspnetcore server integrationtesting selfhostdeployer information dotnet publish stdout copyright c microsoft corporation all rights reserved microsoft aspnetcore server integrationtesting selfhostdeployer information dotnet publish stdout mnt work packages internal aspnetcore sdk build generateassemblyinfo targets error the microsoft aspnetcore buildtools sdk getassemblyfileversion task could not be loaded from the assembly mnt work packages internal aspnetcore sdk build tools internal aspnetcore sdk dll assembly with same name is already loaded confirm that the declaration is correct that the assembly and all its dependencies are available and that the task contains a public class that implements microsoft build framework itask microsoft aspnetcore server integrationtesting selfhostdeployer information dotnet publish stdout mnt work packages internal aspnetcore sdk build generateassemblyinfo targets error the microsoft aspnetcore buildtools sdk getassemblyfileversion task could not be loaded from the assembly mnt work packages internal aspnetcore sdk build tools internal aspnetcore sdk dll assembly with same name is already loaded confirm that the declaration is correct that the assembly and all its dependencies are available and that the task contains a public class that implements microsoft build framework itask microsoft aspnetcore server integrationtesting selfhostdeployer error dotnet publish exited with exit code microsoft aspnetcore server integrationtesting selfhostdeployer warning host process already exited or never started successfully testlifetime information finished test executeshutdowntest in testlifetime information finished test shutdowntestrun in other tests within that build may have failed with a similar message but they are not listed here check the link above for more info this test failed on cc eilon because the bot doesn t know who else to pick this issue was made automatically if there is a problem contact ryanbrandenburg
1
340,814
30,544,906,717
IssuesEvent
2023-07-20 02:26:56
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: schemachange/mixed-versions-compat failed
C-test-failure O-robot O-roachtest T-sql-foundations branch-release-22.2
roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10830317?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10830317?buildTab=artifacts#/schemachange/mixed-versions-compat) on release-22.2 @ [f0bda42400b5bcd7a2ccad3611144793a00f18a8](https://github.com/cockroachdb/cockroach/commits/f0bda42400b5bcd7a2ccad3611144793a00f18a8): ``` test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1 (mixed_version_decl_schemachange_compat.go:64).fetchCorpusToTmpDir: Missing validation corpus for release-22.1 (output in run_142406.749563148_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.1/corpus /tmp/corpus2031648367/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_142406.755148833_n1_gsutil-cp-gscockroac.log: exit status 1) ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29581
2.0
roachtest: schemachange/mixed-versions-compat failed - roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10830317?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10830317?buildTab=artifacts#/schemachange/mixed-versions-compat) on release-22.2 @ [f0bda42400b5bcd7a2ccad3611144793a00f18a8](https://github.com/cockroachdb/cockroach/commits/f0bda42400b5bcd7a2ccad3611144793a00f18a8): ``` test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1 (mixed_version_decl_schemachange_compat.go:64).fetchCorpusToTmpDir: Missing validation corpus for release-22.1 (output in run_142406.749563148_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.1/corpus /tmp/corpus2031648367/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_142406.755148833_n1_gsutil-cp-gscockroac.log: exit status 1) ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29581
test
roachtest schemachange mixed versions compat failed roachtest schemachange mixed versions compat with on release test artifacts and logs in artifacts schemachange mixed versions compat run mixed version decl schemachange compat go fetchcorpustotmpdir missing validation corpus for release output in run gsutil cp gscockroac gsutil cp gs cockroach corpus corpus release corpus tmp corpus returned command problem ssh verbose log retained in ssh gsutil cp gscockroac log exit status parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql foundations jira issue crdb
1
97,828
20,424,198,337
IssuesEvent
2022-02-24 00:50:13
WordPress/openverse-catalog
https://api.github.com/repos/WordPress/openverse-catalog
opened
Enable XCom pickling
good first issue help wanted 🟩 priority: low ✨ goal: improvement 💻 aspect: code 🐳 tech: docker
## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> XComs by default are inserted into the database as strings. While this is more secure, it can make handling native Python types very difficult. ## Description <!-- Describe the feature and how it solves the problem. --> We should [enable XCom pickling](https://airflow.apache.org/docs/apache-airflow/stable/configurations-ref.html#enable-xcom-pickling) in the Airflow configuration. This will make operations which use upstream XComs easier since we will not have to do string/JSON conversions each time we pull the XCom (e.g. see [this comment](https://github.com/WordPress/openverse-catalog/pull/369#discussion_r813215801)). This is something we can add to the [Docker Airflow configuration block in the Dockerfile](https://github.com/WordPress/openverse-catalog/blob/69b2eb7aaa2daf54029fac57b3b99405734da7de/docker/airflow/Dockerfile#L16). ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> [Pickling](https://docs.python.org/3/library/pickle.html) serializes/deserializes native Python objects, which introduces an opportunity for remote code execution. Since our Airflow is behind authentication, I believe that the benefit we gain from pickling is worth this extra risk. In any case, **we should always be careful what data we're passing via XComs**. Those values go into the database, so we need to 1) make sure we trust them and 2) be careful not to try and store too much information in an XCom (e.g. passing Pandas DataFrames via XComs is generally considered a Bad Idea, unless one is certain they will be small). ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
1.0
Enable XCom pickling - ## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> XComs by default are inserted into the database as strings. While this is more secure, it can make handling native Python types very difficult. ## Description <!-- Describe the feature and how it solves the problem. --> We should [enable XCom pickling](https://airflow.apache.org/docs/apache-airflow/stable/configurations-ref.html#enable-xcom-pickling) in the Airflow configuration. This will make operations which use upstream XComs easier since we will not have to do string/JSON conversions each time we pull the XCom (e.g. see [this comment](https://github.com/WordPress/openverse-catalog/pull/369#discussion_r813215801)). This is something we can add to the [Docker Airflow configuration block in the Dockerfile](https://github.com/WordPress/openverse-catalog/blob/69b2eb7aaa2daf54029fac57b3b99405734da7de/docker/airflow/Dockerfile#L16). ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> [Pickling](https://docs.python.org/3/library/pickle.html) serializes/deserializes native Python objects, which introduces an opportunity for remote code execution. Since our Airflow is behind authentication, I believe that the benefit we gain from pickling is worth this extra risk. In any case, **we should always be careful what data we're passing via XComs**. Those values go into the database, so we need to 1) make sure we trust them and 2) be careful not to try and store too much information in an XCom (e.g. passing Pandas DataFrames via XComs is generally considered a Bad Idea, unless one is certain they will be small). ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
non_test
enable xcom pickling problem xcoms by default are inserted into the database as strings while this is more secure it can make handling native python types very difficult description we should in the airflow configuration this will make operations which use upstream xcoms easier since we will not have to do string json conversions each time we pull the xcom e g see this is something we can add to the alternatives additional context serializes deserializes native python objects which introduces an opportunity for remote code execution since our airflow is behind authentication i believe that the benefit we gain from pickling is worth this extra risk in any case we should always be careful what data we re passing via xcoms those values go into the database so we need to make sure we trust them and be careful not to try and store too much information in an xcom e g passing pandas dataframes via xcoms is generally considered a bad idea unless one is certain they will be small implementation 🙋 i would be interested in implementing this feature
0
245,665
20,792,425,615
IssuesEvent
2022-03-17 04:35:33
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql: TestDistSQLRunningInAbortedTxn failed
C-test-failure O-robot branch-master T-sql-queries
sql.TestDistSQLRunningInAbortedTxn [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4575824&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4575824&tab=artifacts#/) on master @ [f5fc84fb5707428ae9505c5e3e90cf3f63d465ad](https://github.com/cockroachdb/cockroach/commits/f5fc84fb5707428ae9505c5e3e90cf3f63d465ad): ``` === RUN TestDistSQLRunningInAbortedTxn test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/233e6aa31ff7ad29c896f0479cb5d0c4/logTestDistSQLRunningInAbortedTxn177929880 test_log_scope.go:80: use -show-logs to present logs inline ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestDistSQLRunningInAbortedTxn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-13789
1.0
sql: TestDistSQLRunningInAbortedTxn failed - sql.TestDistSQLRunningInAbortedTxn [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4575824&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4575824&tab=artifacts#/) on master @ [f5fc84fb5707428ae9505c5e3e90cf3f63d465ad](https://github.com/cockroachdb/cockroach/commits/f5fc84fb5707428ae9505c5e3e90cf3f63d465ad): ``` === RUN TestDistSQLRunningInAbortedTxn test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/233e6aa31ff7ad29c896f0479cb5d0c4/logTestDistSQLRunningInAbortedTxn177929880 test_log_scope.go:80: use -show-logs to present logs inline ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestDistSQLRunningInAbortedTxn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-13789
test
sql testdistsqlrunninginabortedtxn failed sql testdistsqlrunninginabortedtxn with on master run testdistsqlrunninginabortedtxn test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline help see also parameters in this failure tags bazel gss deadlock cc cockroachdb sql queries jira issue crdb
1
83,028
7,859,641,575
IssuesEvent
2018-06-21 17:15:37
coreos/etcd
https://api.github.com/repos/coreos/etcd
closed
etcd integration_e2e test "TestV3CurlCipherSuitesMismatch" is failing
Testing
test log: ``` --- PASS: TestV3CurlCipherSuitesValid (1.12s) === RUN TestV3CurlCipherSuitesMismatch --- FAIL: TestV3CurlCipherSuitesMismatch (1.11s) v3_curl_test.go:391: failed get with curl (read /dev/ptmx: input/output error (expected "alert handshake failure", got ["# HELP etcd_debugging_lease_granted_total The total number of granted leases. # TYPE etcd_debugging_lease_granted_total counter etcd_debugging_lease_granted_total # HELP etcd_debugging_lease_renewed_total The number of renewed leases seen by the leader. # TYPE etcd_debugging_lease_renewed_total counter etcd_debugging_lease_renewed_total # HELP etcd_debugging_lease_revoked_total The total number of revoked leases. # TYPE etcd_debugging_lease_revoked_total counter etcd_debugging_lease_revoked_total # HELP etcd_debugging_lease_ttl_total Bucketed histogram of lease TTLs. # TYPE etcd_debugging_lease_ttl_total histogram etcd_debugging_lease_ttl_total_bucket{le=\"1\"} etcd_debugging_lease_ttl_total_bucket{le=\"2\"} etcd_debugging_lease_ttl_total_bucket{le=\"4\"} etcd_debugging_lease_ttl_total_bucket{le=\"8\"} etcd_debugging_lease_ttl_total_bucket{le=\"16\"} etcd_debugging_lease_ttl_total_bucket{le=\"32\"} etcd_debugging_lease_ttl_total_bucket{le=\"64\"} etcd_debugging_lease_ttl_total_bucket{le=\"128\"} etcd_debugging_lease_ttl_total_bucket{le=\"256\"} etcd_debugging_lease_ttl_total_bucket{le=\"512\"} etcd_debugging_lease_ttl_total_bucket{le=\"1024\"} etcd_debugging_lease_ttl_total_bucket{le=\"2048\"} etcd_debugging_lease_ttl_total_bucket{le=\"4096\"} etcd_debugging_lease_ttl_total_bucket{le=\"8192\"} etcd_debugging_lease_ttl_total_bucket{le=\"16384\"} etcd_debugging_lease_ttl_total_bucket{le=\"32768\"} etcd_debugging_lease_ttl_total_bucket{le=\"65536\"} etcd_debugging_lease_ttl_total_bucket{le=\"131072\"} etcd_debugging_lease_ttl_total_bucket{le=\"262144\"} etcd_debugging_lease_ttl_total_bucket{le=\"524288\"} etcd_debugging_lease_ttl_total_bucket{le=\"1.048576e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"2.097152e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"4.194304e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"8.388608e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"+Inf\"} etcd_debugging_lease_ttl_total_sum etcd_debugging_lease_ttl_total_count # HELP etcd_debugging_mvcc_db_compaction_keys_total Total number of db keys compacted. # TYPE etcd_debugging_mvcc_db_compaction_keys_total counter etcd_debugging_mvcc_db_compaction_keys_total # HELP etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds Bucketed histogram of db compaction pause duration. # TYPE etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds histogram etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"1\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"2\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"4\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"8\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"16\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"32\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"64\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"128\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"256\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"512\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"1024\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"2048\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"4096\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_sum etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_count # HELP etcd_debugging_mvcc_db_compaction_total_duration_milliseconds Bucketed histogram of db compaction total duration. # TYPE etcd_debugging_mvcc_db_compaction_total_duration_milliseconds histogram etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"100\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"1600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"3200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"6400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"12800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"25600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"51200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"102400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"204800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"409600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"819200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_sum etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_count # HELP etcd_debugging_mvcc_db_total_size_in_bytes Total size of the underlying database physically allocated in bytes. # TYPE etcd_debugging_mvcc_db_total_size_in_bytes gauge etcd_debugging_mvcc_db_total_size_in_bytes 32768 # HELP etcd_debugging_mvcc_delete_total Total number of deletes seen by this member. # TYPE etcd_debugging_mvcc_delete_total counter etcd_debugging_mvcc_delete_total # HELP etcd_debugging_mvcc_events_total Total number of events sent by this member. # TYPE etcd_debugging_mvcc_events_total counter etcd_debugging_mvcc_events_total # HELP etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds Bucketed histogram of index compaction pause duration. # TYPE etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds histogram etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"0.5\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"1\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"2\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"4\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"8\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"16\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"32\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"64\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"128\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"256\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"512\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"1024\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"2048\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"4096\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_sum etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_count # HELP etcd_debugging_mvcc_keys_total Total number of keys. # TYPE etcd_debugging_mvcc_keys_total gauge etcd_debugging_mvcc_keys_total # HELP etcd_debugging_mvcc_pending_events_total Total number of pending events to be sent. # TYPE etcd_debugging_mvcc_pending_events_total gauge etcd_debugging_mvcc_pending_events_total # HELP etcd_debugging_mvcc_put_total Total number of puts seen by this member. # TYPE etcd_debugging_mvcc_put_total counter etcd_debugging_mvcc_put_total # HELP etcd_debugging_mvcc_range_total Total number of ranges seen by this member. # TYPE etcd_debugging_mvcc_range_total counter etcd_debugging_mvcc_range_total # HELP etcd_debugging_mvcc_slow_watcher_total Total number of unsynced slow watchers. # TYPE etcd_debugging_mvcc_slow_watcher_total gauge etcd_debugging_mvcc_slow_watcher_total # HELP etcd_debugging_mvcc_txn_total Total number of txns seen by this member. # TYPE etcd_debugging_mvcc_txn_total counter etcd_debugging_mvcc_txn_total # HELP etcd_debugging_mvcc_watch_stream_total Total number of watch streams. # TYPE etcd_debugging_mvcc_watch_stream_total gauge etcd_debugging_mvcc_watch_stream_total # HELP etcd_debugging_mvcc_watcher_total Total number of watchers. # TYPE etcd_debugging_mvcc_watcher_total gauge etcd_debugging_mvcc_watcher_total # HELP etcd_debugging_server_lease_expired_total The total number of expired leases. # TYPE etcd_debugging_server_lease_expired_total counter etcd_debugging_server_lease_expired_total # HELP etcd_debugging_snap_save_marshalling_duration_seconds The marshalling cost distributions of save called by snapshot. # TYPE etcd_debugging_snap_save_marshalling_duration_seconds histogram etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.001\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.002\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.004\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.008\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.016\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.032\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.064\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.128\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.256\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.512\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"1.024\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"2.048\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"4.096\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"8.192\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"+Inf\"} etcd_debugging_snap_save_marshalling_duration_seconds_sum etcd_debugging_snap_save_marshalling_duration_seconds_count # HELP etcd_debugging_snap_save_total_duration_seconds The total latency distributions of save called by snapshot. # TYPE etcd_debugging_snap_save_total_duration_seconds histogram etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.001\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.002\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.004\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.008\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.016\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.032\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.064\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.128\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.256\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.512\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"1.024\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"2.048\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"4.096\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"8.192\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"+Inf\"} etcd_debugging_snap_save_total_duration_seconds_sum etcd_debugging_snap_save_total_duration_seconds_count # HELP etcd_debugging_store_expires_total Total number of expired keys. # TYPE etcd_debugging_store_expires_total counter etcd_debugging_store_expires_total # HELP etcd_debugging_store_reads_total Total number of reads action by (get/getRecursive), local to this member. # TYPE etcd_debugging_store_reads_total counter etcd_debugging_store_reads_total{action=\"getRecursive\"} 1 # HELP etcd_debugging_store_watch_requests_total Total number of incoming watch requests (new or reestablished). # TYPE etcd_debugging_store_watch_requests_total counter etcd_debugging_store_watch_requests_total # HELP etcd_debugging_store_watchers Count of currently active watchers. # TYPE etcd_debugging_store_watchers gauge etcd_debugging_store_watchers # HELP etcd_debugging_store_writes_total Total number of writes (e.g. set/compareAndDelete) seen by this member. # TYPE etcd_debugging_store_writes_total counter etcd_debugging_store_writes_total{action=\"create\"} 1 etcd_debugging_store_writes_total{action=\"set\"} 2 # HELP etcd_disk_backend_commit_duration_seconds The latency distributions of commit called by backend. # TYPE etcd_disk_backend_commit_duration_seconds histogram etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.001\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.002\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.004\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.008\"} 4 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.016\"} 5 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.032\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.064\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.128\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.256\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.512\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"1.024\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"2.048\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"4.096\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"8.192\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"+Inf\"} 6 etcd_disk_backend_commit_duration_seconds_sum 0.061365869999999996 etcd_disk_backend_commit_duration_seconds_count 6 # HELP etcd_disk_backend_defrag_duration_seconds The latency distribution of backend defragmentation. # TYPE etcd_disk_backend_defrag_duration_seconds histogram etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.1\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"1.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"3.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"6.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"12.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"25.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"51.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"102.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"204.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"409.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"+Inf\"} etcd_disk_backend_defrag_duration_seconds_sum etcd_disk_backend_defrag_duration_seconds_count # HELP etcd_disk_backend_snapshot_duration_seconds The latency distribution of backend snapshots. # TYPE etcd_disk_backend_snapshot_duration_seconds histogram etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.01\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.02\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.04\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.08\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.16\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.32\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.64\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"1.28\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"2.56\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"5.12\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"10.24\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"20.48\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"40.96\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"81.92\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"163.84\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"327.68\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"655.36\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"+Inf\"} etcd_disk_backend_snapshot_duration_seconds_sum etcd_disk_backend_snapshot_duration_seconds_count # HELP etcd_disk_wal_fsync_duration_seconds The latency distributions of fsync called by WAL. # TYPE etcd_disk_wal_fsync_duration_seconds histogram etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.001\"} 1 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.002\"} 2 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.004\"} 3 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.008\"} 3 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.016\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.032\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.064\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.128\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.256\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.512\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"1.024\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"2.048\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"4.096\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"8.192\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"+Inf\"} 4 etcd_disk_wal_fsync_duration_seconds_sum 0.012822563 etcd_disk_wal_fsync_duration_seconds_count 4 # HELP etcd_grpc_proxy_cache_hits_total Total number of cache hits # TYPE etcd_grpc_proxy_cache_hits_total gauge etcd_grpc_proxy_cache_hits_total # HELP etcd_grpc_proxy_cache_keys_total Total number of keys/ranges cached # TYPE etcd_grpc_proxy_cache_keys_total gauge etcd_grpc_proxy_cache_keys_total # HELP etcd_grpc_proxy_cache_misses_total Total number of cache misses # TYPE etcd_grpc_proxy_cache_misses_total gauge etcd_grpc_proxy_cache_misses_total # HELP etcd_grpc_proxy_events_coalescing_total Total number of events coalescing # TYPE etcd_grpc_proxy_events_coalescing_total counter etcd_grpc_proxy_events_coalescing_total # HELP etcd_grpc_proxy_watchers_coalescing_total Total number of current watchers coalescing # TYPE etcd_grpc_proxy_watchers_coalescing_total gauge etcd_grpc_proxy_watchers_coalescing_total # HELP etcd_mvcc_db_total_size_in_bytes Total size of the underlying database physically allocated in bytes. # TYPE etcd_mvcc_db_total_size_in_bytes gauge etcd_mvcc_db_total_size_in_bytes 32768 # HELP etcd_mvcc_db_total_size_in_use_in_bytes Total size of the underlying database logically in use in bytes. # TYPE etcd_mvcc_db_total_size_in_use_in_bytes gauge etcd_mvcc_db_total_size_in_use_in_bytes 262144 # HELP etcd_mvcc_hash_duration_seconds The latency distribution of storage hash operation. # TYPE etcd_mvcc_hash_duration_seconds histogram etcd_mvcc_hash_duration_seconds_bucket{le=\"0.01\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.02\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.04\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.08\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.16\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.32\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.64\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"1.28\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"2.56\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"5.12\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"10.24\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"20.48\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"40.96\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"81.92\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"163.84\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"+Inf\"} etcd_mvcc_hash_duration_seconds_sum etcd_mvcc_hash_duration_seconds_count # HELP etcd_mvcc_hash_rev_duration_seconds The latency distribution of storage hash by revision operation. # TYPE etcd_mvcc_hash_rev_duration_seconds histogram etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.01\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.02\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.04\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.08\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.16\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.32\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.64\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"1.28\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"2.56\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"5.12\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"10.24\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"20.48\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"40.96\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"81.92\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"163.84\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"+Inf\"} etcd_mvcc_hash_rev_duration_seconds_sum etcd_mvcc_hash_rev_duration_seconds_count # HELP etcd_network_client_grpc_received_bytes_total The total number of bytes received from grpc clients. # TYPE etcd_network_client_grpc_received_bytes_total counter etcd_network_client_grpc_received_bytes_total # HELP etcd_network_client_grpc_sent_bytes_total The total number of bytes sent to grpc clients. # TYPE etcd_network_client_grpc_sent_bytes_total counter etcd_network_client_grpc_sent_bytes_total # HELP etcd_server_has_leader Whether or not a leader exists. 1 is existence, 0 is not. # TYPE etcd_server_has_leader gauge etcd_server_has_leader 1 # HELP etcd_server_heartbeat_send_failures_total The total number of leader heartbeat send failures (likely overloaded from slow disk). # TYPE etcd_server_heartbeat_send_failures_total counter etcd_server_heartbeat_send_failures_total # HELP etcd_server_is_leader Whether or not this member is a leader. 1 if is, 0 otherwise. # TYPE etcd_server_is_leader gauge etcd_server_is_leader 1 # HELP etcd_server_leader_changes_seen_total The number of leader changes seen. # TYPE etcd_server_leader_changes_seen_total counter etcd_server_leader_changes_seen_total 1 # HELP etcd_server_proposals_applied_total The total number of consensus proposals applied. # TYPE etcd_server_proposals_applied_total gauge etcd_server_proposals_applied_total 4 # HELP etcd_server_proposals_committed_total The total number of consensus proposals committed. # TYPE etcd_server_proposals_committed_total gauge etcd_server_proposals_committed_total 4 # HELP etcd_server_proposals_failed_total The total number of failed proposals seen. # TYPE etcd_server_proposals_failed_total counter etcd_server_proposals_failed_total # HELP etcd_server_proposals_pending The current number of pending proposals to commit. # TYPE etcd_server_proposals_pending gauge etcd_server_proposals_pending # HELP etcd_server_quota_backend_bytes Current backend storage quota size in bytes. # TYPE etcd_server_quota_backend_bytes gauge etcd_server_quota_backend_bytes 2.147483648e+09 # HELP etcd_server_slow_apply_total The total number of slow apply requests (likely overloaded from slow disk). # TYPE etcd_server_slow_apply_total counter etcd_server_slow_apply_total # HELP etcd_server_version Which version is running. 1 for 'server_version' label with current version. # TYPE etcd_server_version gauge etcd_server_version{server_version=\"3.3.0+git\"} 1 # HELP etcd_snap_fsync_duration_seconds The latency distributions of fsync called by snap. # TYPE etcd_snap_fsync_duration_seconds histogram etcd_snap_fsync_duration_seconds_bucket{le=\"0.001\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.002\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.004\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.008\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.016\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.032\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.064\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.128\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.256\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.512\"} etcd_snap_fsync_duration_seconds_bucket{le=\"1.024\"} etcd_snap_fsync_duration_seconds_bucket{le=\"2.048\"} etcd_snap_fsync_duration_seconds_bucket{le=\"4.096\"} etcd_snap_fsync_duration_seconds_bucket{le=\"8.192\"} etcd_snap_fsync_duration_seconds_bucket{le=\"+Inf\"} etcd_snap_fsync_duration_seconds_sum etcd_snap_fsync_duration_seconds_count # HELP go_gc_duration_seconds A summary of the GC invocation durations. # TYPE go_gc_duration_seconds summary go_gc_duration_seconds{quantile=\"0\"} 0.00015764 go_gc_duration_seconds{quantile=\"0.25\"} 0.00015764 go_gc_duration_seconds{quantile=\"0.5\"} 0.000177495 go_gc_duration_seconds{quantile=\"0.75\"} 0.000184464 go_gc_duration_seconds{quantile=\"1\"} 0.000184464 go_gc_duration_seconds_sum 0.000519599 go_gc_duration_seconds_count 3 # HELP go_goroutines Number of goroutines that currently exist. # TYPE go_goroutines gauge go_goroutines 47 # HELP go_info Information about the Go environment. # TYPE go_info gauge go_info{version=\"devel +b74f732 Mon Jun 11 00:45:20 2018 +0000\"} 1 # HELP go_memstats_alloc_bytes Number of bytes allocated and still in use. # TYPE go_memstats_alloc_bytes gauge go_memstats_alloc_bytes 4.150712e+06 # HELP go_memstats_alloc_bytes_total Total number of bytes allocated, even if freed. # TYPE go_memstats_alloc_bytes_total counter go_memstats_alloc_bytes_total 8.208744e+06 # HELP go_memstats_buck_hash_sys_bytes Number of bytes used by the profiling bucket hash table. # TYPE go_memstats_buck_hash_sys_bytes gauge go_memstats_buck_hash_sys_bytes 1.446981e+06 # HELP go_memstats_frees_total Total number of frees. # TYPE go_memstats_frees_total counter go_memstats_frees_total 996 # HELP go_memstats_gc_cpu_fraction The fraction of this program's available CPU time used by the GC since the program started. # TYPE go_memstats_gc_cpu_fraction gauge go_memstats_gc_cpu_fraction 0.001566909347503071 # HELP go_memstats_gc_sys_bytes Number of bytes used for garbage collection system metadata. # TYPE go_memstats_gc_sys_bytes gauge go_memstats_gc_sys_bytes 2.371584e+06 # HELP go_memstats_heap_alloc_bytes Number of heap bytes allocated and still in use. # TYPE go_memstats_heap_alloc_bytes gauge go_memstats_heap_alloc_bytes 4.150712e+06 # HELP go_memstats_heap_idle_bytes Number of heap bytes waiting to be used. # TYPE go_memstats_heap_idle_bytes gauge go_memstats_heap_idle_bytes 5.9342848e+07 # HELP go_memstats_heap_inuse_bytes Number of heap bytes that are in use. # TYPE go_memstats_heap_inuse_bytes gauge go_memstats_heap_inuse_bytes 6.22592e+06 # HELP go_memstats_heap_objects Number of allocated objects. # TYPE go_memstats_heap_objects gauge go_memstats_heap_objects 2812 # HELP go_memstats_heap_released_bytes Number of heap bytes released to OS. # TYPE go_memstats_heap_released_bytes gauge go_memstats_heap_released_bytes # HELP go_memstats_heap_sys_bytes Number of heap bytes obtained from system. # TYPE go_memstats_heap_sys_bytes gauge go_memstats_heap_sys_bytes 6.5568768e+07 # HELP go_memstats_last_gc_time_seconds Number of seconds since 1970 of last garbage collection. # TYPE go_memstats_last_gc_time_seconds gauge go_memstats_last_gc_time_seconds 1.5286935487193735e+09 # HELP go_memstats_lookups_total Total number of pointer lookups. # TYPE go_memstats_lookups_total counter go_memstats_lookups_total # HELP go_memstats_mallocs_total Total number of mallocs. # TYPE go_memstats_mallocs_total counter go_memstats_mallocs_total 3808 # HELP go_memstats_mcache_inuse_bytes Number of bytes in use by mcache structures. # TYPE go_memstats_mcache_inuse_bytes gauge go_memstats_mcache_inuse_bytes 27648 # HELP go_memstats_mcache_sys_bytes Number of bytes used for mcache structures obtained from system. # TYPE go_memstats_mcache_sys_bytes gauge go_memstats_mcache_sys_bytes 32768 # HELP go_memstats_mspan_inuse_bytes Number of bytes in use by mspan structures. # TYPE go_memstats_mspan_inuse_bytes gauge go_memstats_mspan_inuse_bytes 74784 # HELP go_memstats_mspan_sys_bytes Number of bytes used for mspan structures obtained from system. # TYPE go_memstats_mspan_sys_bytes gauge go_memstats_mspan_sys_bytes 8192 # HELP go_memstats_next_gc_bytes Number of heap bytes when next garbage collection will take place. # TYPE go_memstats_next_gc_bytes gauge go_memstats_next_gc_bytes 6.983584e+06 # HELP go_memstats_other_sys_bytes Number of bytes used for other system allocations. # TYPE go_memstats_other_sys_bytes gauge go_memstats_other_sys_bytes 2.290867e+06 # HELP go_memstats_stack_inuse_bytes Number of bytes in use by the stack allocator. # TYPE go_memstats_stack_inuse_bytes gauge go_memstats_stack_inuse_bytes 1.540096e+06 # HELP go_memstats_stack_sys_bytes Number of bytes obtained from system for stack allocator. # TYPE go_memstats_stack_sys_bytes gauge go_memstats_stack_sys_bytes 1.540096e+06 # HELP go_memstats_sys_bytes Number of bytes obtained from system. # TYPE go_memstats_sys_bytes gauge go_memstats_sys_bytes 7.3332984e+07 # HELP go_threads Number of OS threads created. # TYPE go_threads gauge go_threads 23 # HELP grpc_server_handled_total Total number of RPCs completed on the server, regardless of success or failure. # TYPE grpc_server_handled_total counter grpc_server_handled_total{grpc_code=\"Aborted\",grpc_method=\"Alarm\",grpc_service=\"etcdserverpb.Maintenance\",grpc_type=\"unary\"} grpc_server_handled_total{grpc_code=\"Aborted\",grpc_method=\"AuthDisable\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} <snip> grpc_server_started_total{grpc_method=\"UserList\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} grpc_server_started_total{grpc_method=\"UserRevokeRole\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} grpc_server_started_total{grpc_method=\"Watch\",grpc_service=\"etcdserverpb.Watch\",grpc_type=\"bidi_stream\"} # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. # TYPE process_cpu_seconds_total counter process_cpu_seconds_total 0.12 # HELP process_max_fds Maximum number of open file descriptors. # TYPE process_max_fds gauge process_max_fds 1.048576e+06 # HELP process_open_fds Number of open file descriptors. # TYPE process_open_fds gauge process_open_fds 14 # HELP process_resident_memory_bytes Resident memory size in bytes. # TYPE process_resident_memory_bytes gauge process_resident_memory_bytes 1.7301504e+07 # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. # TYPE process_start_time_seconds gauge process_start_time_seconds 1.52869354775e+09 # HELP process_virtual_memory_bytes Virtual memory size in bytes. # TYPE process_virtual_memory_bytes gauge process_virtual_memory_bytes 1.08675072e+1 "])) === RUN TestCtlV2Set sudeeshjohn:/tmp$ ``` etcd version: 3.3.0+git
1.0
etcd integration_e2e test "TestV3CurlCipherSuitesMismatch" is failing - test log: ``` --- PASS: TestV3CurlCipherSuitesValid (1.12s) === RUN TestV3CurlCipherSuitesMismatch --- FAIL: TestV3CurlCipherSuitesMismatch (1.11s) v3_curl_test.go:391: failed get with curl (read /dev/ptmx: input/output error (expected "alert handshake failure", got ["# HELP etcd_debugging_lease_granted_total The total number of granted leases. # TYPE etcd_debugging_lease_granted_total counter etcd_debugging_lease_granted_total # HELP etcd_debugging_lease_renewed_total The number of renewed leases seen by the leader. # TYPE etcd_debugging_lease_renewed_total counter etcd_debugging_lease_renewed_total # HELP etcd_debugging_lease_revoked_total The total number of revoked leases. # TYPE etcd_debugging_lease_revoked_total counter etcd_debugging_lease_revoked_total # HELP etcd_debugging_lease_ttl_total Bucketed histogram of lease TTLs. # TYPE etcd_debugging_lease_ttl_total histogram etcd_debugging_lease_ttl_total_bucket{le=\"1\"} etcd_debugging_lease_ttl_total_bucket{le=\"2\"} etcd_debugging_lease_ttl_total_bucket{le=\"4\"} etcd_debugging_lease_ttl_total_bucket{le=\"8\"} etcd_debugging_lease_ttl_total_bucket{le=\"16\"} etcd_debugging_lease_ttl_total_bucket{le=\"32\"} etcd_debugging_lease_ttl_total_bucket{le=\"64\"} etcd_debugging_lease_ttl_total_bucket{le=\"128\"} etcd_debugging_lease_ttl_total_bucket{le=\"256\"} etcd_debugging_lease_ttl_total_bucket{le=\"512\"} etcd_debugging_lease_ttl_total_bucket{le=\"1024\"} etcd_debugging_lease_ttl_total_bucket{le=\"2048\"} etcd_debugging_lease_ttl_total_bucket{le=\"4096\"} etcd_debugging_lease_ttl_total_bucket{le=\"8192\"} etcd_debugging_lease_ttl_total_bucket{le=\"16384\"} etcd_debugging_lease_ttl_total_bucket{le=\"32768\"} etcd_debugging_lease_ttl_total_bucket{le=\"65536\"} etcd_debugging_lease_ttl_total_bucket{le=\"131072\"} etcd_debugging_lease_ttl_total_bucket{le=\"262144\"} etcd_debugging_lease_ttl_total_bucket{le=\"524288\"} etcd_debugging_lease_ttl_total_bucket{le=\"1.048576e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"2.097152e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"4.194304e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"8.388608e+06\"} etcd_debugging_lease_ttl_total_bucket{le=\"+Inf\"} etcd_debugging_lease_ttl_total_sum etcd_debugging_lease_ttl_total_count # HELP etcd_debugging_mvcc_db_compaction_keys_total Total number of db keys compacted. # TYPE etcd_debugging_mvcc_db_compaction_keys_total counter etcd_debugging_mvcc_db_compaction_keys_total # HELP etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds Bucketed histogram of db compaction pause duration. # TYPE etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds histogram etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"1\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"2\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"4\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"8\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"16\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"32\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"64\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"128\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"256\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"512\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"1024\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"2048\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"4096\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_sum etcd_debugging_mvcc_db_compaction_pause_duration_milliseconds_count # HELP etcd_debugging_mvcc_db_compaction_total_duration_milliseconds Bucketed histogram of db compaction total duration. # TYPE etcd_debugging_mvcc_db_compaction_total_duration_milliseconds histogram etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"100\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"1600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"3200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"6400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"12800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"25600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"51200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"102400\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"204800\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"409600\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"819200\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_sum etcd_debugging_mvcc_db_compaction_total_duration_milliseconds_count # HELP etcd_debugging_mvcc_db_total_size_in_bytes Total size of the underlying database physically allocated in bytes. # TYPE etcd_debugging_mvcc_db_total_size_in_bytes gauge etcd_debugging_mvcc_db_total_size_in_bytes 32768 # HELP etcd_debugging_mvcc_delete_total Total number of deletes seen by this member. # TYPE etcd_debugging_mvcc_delete_total counter etcd_debugging_mvcc_delete_total # HELP etcd_debugging_mvcc_events_total Total number of events sent by this member. # TYPE etcd_debugging_mvcc_events_total counter etcd_debugging_mvcc_events_total # HELP etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds Bucketed histogram of index compaction pause duration. # TYPE etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds histogram etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"0.5\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"1\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"2\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"4\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"8\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"16\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"32\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"64\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"128\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"256\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"512\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"1024\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"2048\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"4096\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_bucket{le=\"+Inf\"} etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_sum etcd_debugging_mvcc_index_compaction_pause_duration_milliseconds_count # HELP etcd_debugging_mvcc_keys_total Total number of keys. # TYPE etcd_debugging_mvcc_keys_total gauge etcd_debugging_mvcc_keys_total # HELP etcd_debugging_mvcc_pending_events_total Total number of pending events to be sent. # TYPE etcd_debugging_mvcc_pending_events_total gauge etcd_debugging_mvcc_pending_events_total # HELP etcd_debugging_mvcc_put_total Total number of puts seen by this member. # TYPE etcd_debugging_mvcc_put_total counter etcd_debugging_mvcc_put_total # HELP etcd_debugging_mvcc_range_total Total number of ranges seen by this member. # TYPE etcd_debugging_mvcc_range_total counter etcd_debugging_mvcc_range_total # HELP etcd_debugging_mvcc_slow_watcher_total Total number of unsynced slow watchers. # TYPE etcd_debugging_mvcc_slow_watcher_total gauge etcd_debugging_mvcc_slow_watcher_total # HELP etcd_debugging_mvcc_txn_total Total number of txns seen by this member. # TYPE etcd_debugging_mvcc_txn_total counter etcd_debugging_mvcc_txn_total # HELP etcd_debugging_mvcc_watch_stream_total Total number of watch streams. # TYPE etcd_debugging_mvcc_watch_stream_total gauge etcd_debugging_mvcc_watch_stream_total # HELP etcd_debugging_mvcc_watcher_total Total number of watchers. # TYPE etcd_debugging_mvcc_watcher_total gauge etcd_debugging_mvcc_watcher_total # HELP etcd_debugging_server_lease_expired_total The total number of expired leases. # TYPE etcd_debugging_server_lease_expired_total counter etcd_debugging_server_lease_expired_total # HELP etcd_debugging_snap_save_marshalling_duration_seconds The marshalling cost distributions of save called by snapshot. # TYPE etcd_debugging_snap_save_marshalling_duration_seconds histogram etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.001\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.002\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.004\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.008\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.016\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.032\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.064\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.128\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.256\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"0.512\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"1.024\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"2.048\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"4.096\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"8.192\"} etcd_debugging_snap_save_marshalling_duration_seconds_bucket{le=\"+Inf\"} etcd_debugging_snap_save_marshalling_duration_seconds_sum etcd_debugging_snap_save_marshalling_duration_seconds_count # HELP etcd_debugging_snap_save_total_duration_seconds The total latency distributions of save called by snapshot. # TYPE etcd_debugging_snap_save_total_duration_seconds histogram etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.001\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.002\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.004\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.008\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.016\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.032\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.064\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.128\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.256\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"0.512\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"1.024\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"2.048\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"4.096\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"8.192\"} etcd_debugging_snap_save_total_duration_seconds_bucket{le=\"+Inf\"} etcd_debugging_snap_save_total_duration_seconds_sum etcd_debugging_snap_save_total_duration_seconds_count # HELP etcd_debugging_store_expires_total Total number of expired keys. # TYPE etcd_debugging_store_expires_total counter etcd_debugging_store_expires_total # HELP etcd_debugging_store_reads_total Total number of reads action by (get/getRecursive), local to this member. # TYPE etcd_debugging_store_reads_total counter etcd_debugging_store_reads_total{action=\"getRecursive\"} 1 # HELP etcd_debugging_store_watch_requests_total Total number of incoming watch requests (new or reestablished). # TYPE etcd_debugging_store_watch_requests_total counter etcd_debugging_store_watch_requests_total # HELP etcd_debugging_store_watchers Count of currently active watchers. # TYPE etcd_debugging_store_watchers gauge etcd_debugging_store_watchers # HELP etcd_debugging_store_writes_total Total number of writes (e.g. set/compareAndDelete) seen by this member. # TYPE etcd_debugging_store_writes_total counter etcd_debugging_store_writes_total{action=\"create\"} 1 etcd_debugging_store_writes_total{action=\"set\"} 2 # HELP etcd_disk_backend_commit_duration_seconds The latency distributions of commit called by backend. # TYPE etcd_disk_backend_commit_duration_seconds histogram etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.001\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.002\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.004\"} etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.008\"} 4 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.016\"} 5 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.032\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.064\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.128\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.256\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"0.512\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"1.024\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"2.048\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"4.096\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"8.192\"} 6 etcd_disk_backend_commit_duration_seconds_bucket{le=\"+Inf\"} 6 etcd_disk_backend_commit_duration_seconds_sum 0.061365869999999996 etcd_disk_backend_commit_duration_seconds_count 6 # HELP etcd_disk_backend_defrag_duration_seconds The latency distribution of backend defragmentation. # TYPE etcd_disk_backend_defrag_duration_seconds histogram etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.1\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"0.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"1.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"3.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"6.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"12.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"25.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"51.2\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"102.4\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"204.8\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"409.6\"} etcd_disk_backend_defrag_duration_seconds_bucket{le=\"+Inf\"} etcd_disk_backend_defrag_duration_seconds_sum etcd_disk_backend_defrag_duration_seconds_count # HELP etcd_disk_backend_snapshot_duration_seconds The latency distribution of backend snapshots. # TYPE etcd_disk_backend_snapshot_duration_seconds histogram etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.01\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.02\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.04\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.08\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.16\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.32\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"0.64\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"1.28\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"2.56\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"5.12\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"10.24\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"20.48\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"40.96\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"81.92\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"163.84\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"327.68\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"655.36\"} etcd_disk_backend_snapshot_duration_seconds_bucket{le=\"+Inf\"} etcd_disk_backend_snapshot_duration_seconds_sum etcd_disk_backend_snapshot_duration_seconds_count # HELP etcd_disk_wal_fsync_duration_seconds The latency distributions of fsync called by WAL. # TYPE etcd_disk_wal_fsync_duration_seconds histogram etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.001\"} 1 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.002\"} 2 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.004\"} 3 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.008\"} 3 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.016\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.032\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.064\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.128\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.256\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"0.512\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"1.024\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"2.048\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"4.096\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"8.192\"} 4 etcd_disk_wal_fsync_duration_seconds_bucket{le=\"+Inf\"} 4 etcd_disk_wal_fsync_duration_seconds_sum 0.012822563 etcd_disk_wal_fsync_duration_seconds_count 4 # HELP etcd_grpc_proxy_cache_hits_total Total number of cache hits # TYPE etcd_grpc_proxy_cache_hits_total gauge etcd_grpc_proxy_cache_hits_total # HELP etcd_grpc_proxy_cache_keys_total Total number of keys/ranges cached # TYPE etcd_grpc_proxy_cache_keys_total gauge etcd_grpc_proxy_cache_keys_total # HELP etcd_grpc_proxy_cache_misses_total Total number of cache misses # TYPE etcd_grpc_proxy_cache_misses_total gauge etcd_grpc_proxy_cache_misses_total # HELP etcd_grpc_proxy_events_coalescing_total Total number of events coalescing # TYPE etcd_grpc_proxy_events_coalescing_total counter etcd_grpc_proxy_events_coalescing_total # HELP etcd_grpc_proxy_watchers_coalescing_total Total number of current watchers coalescing # TYPE etcd_grpc_proxy_watchers_coalescing_total gauge etcd_grpc_proxy_watchers_coalescing_total # HELP etcd_mvcc_db_total_size_in_bytes Total size of the underlying database physically allocated in bytes. # TYPE etcd_mvcc_db_total_size_in_bytes gauge etcd_mvcc_db_total_size_in_bytes 32768 # HELP etcd_mvcc_db_total_size_in_use_in_bytes Total size of the underlying database logically in use in bytes. # TYPE etcd_mvcc_db_total_size_in_use_in_bytes gauge etcd_mvcc_db_total_size_in_use_in_bytes 262144 # HELP etcd_mvcc_hash_duration_seconds The latency distribution of storage hash operation. # TYPE etcd_mvcc_hash_duration_seconds histogram etcd_mvcc_hash_duration_seconds_bucket{le=\"0.01\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.02\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.04\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.08\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.16\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.32\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"0.64\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"1.28\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"2.56\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"5.12\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"10.24\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"20.48\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"40.96\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"81.92\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"163.84\"} etcd_mvcc_hash_duration_seconds_bucket{le=\"+Inf\"} etcd_mvcc_hash_duration_seconds_sum etcd_mvcc_hash_duration_seconds_count # HELP etcd_mvcc_hash_rev_duration_seconds The latency distribution of storage hash by revision operation. # TYPE etcd_mvcc_hash_rev_duration_seconds histogram etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.01\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.02\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.04\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.08\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.16\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.32\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"0.64\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"1.28\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"2.56\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"5.12\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"10.24\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"20.48\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"40.96\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"81.92\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"163.84\"} etcd_mvcc_hash_rev_duration_seconds_bucket{le=\"+Inf\"} etcd_mvcc_hash_rev_duration_seconds_sum etcd_mvcc_hash_rev_duration_seconds_count # HELP etcd_network_client_grpc_received_bytes_total The total number of bytes received from grpc clients. # TYPE etcd_network_client_grpc_received_bytes_total counter etcd_network_client_grpc_received_bytes_total # HELP etcd_network_client_grpc_sent_bytes_total The total number of bytes sent to grpc clients. # TYPE etcd_network_client_grpc_sent_bytes_total counter etcd_network_client_grpc_sent_bytes_total # HELP etcd_server_has_leader Whether or not a leader exists. 1 is existence, 0 is not. # TYPE etcd_server_has_leader gauge etcd_server_has_leader 1 # HELP etcd_server_heartbeat_send_failures_total The total number of leader heartbeat send failures (likely overloaded from slow disk). # TYPE etcd_server_heartbeat_send_failures_total counter etcd_server_heartbeat_send_failures_total # HELP etcd_server_is_leader Whether or not this member is a leader. 1 if is, 0 otherwise. # TYPE etcd_server_is_leader gauge etcd_server_is_leader 1 # HELP etcd_server_leader_changes_seen_total The number of leader changes seen. # TYPE etcd_server_leader_changes_seen_total counter etcd_server_leader_changes_seen_total 1 # HELP etcd_server_proposals_applied_total The total number of consensus proposals applied. # TYPE etcd_server_proposals_applied_total gauge etcd_server_proposals_applied_total 4 # HELP etcd_server_proposals_committed_total The total number of consensus proposals committed. # TYPE etcd_server_proposals_committed_total gauge etcd_server_proposals_committed_total 4 # HELP etcd_server_proposals_failed_total The total number of failed proposals seen. # TYPE etcd_server_proposals_failed_total counter etcd_server_proposals_failed_total # HELP etcd_server_proposals_pending The current number of pending proposals to commit. # TYPE etcd_server_proposals_pending gauge etcd_server_proposals_pending # HELP etcd_server_quota_backend_bytes Current backend storage quota size in bytes. # TYPE etcd_server_quota_backend_bytes gauge etcd_server_quota_backend_bytes 2.147483648e+09 # HELP etcd_server_slow_apply_total The total number of slow apply requests (likely overloaded from slow disk). # TYPE etcd_server_slow_apply_total counter etcd_server_slow_apply_total # HELP etcd_server_version Which version is running. 1 for 'server_version' label with current version. # TYPE etcd_server_version gauge etcd_server_version{server_version=\"3.3.0+git\"} 1 # HELP etcd_snap_fsync_duration_seconds The latency distributions of fsync called by snap. # TYPE etcd_snap_fsync_duration_seconds histogram etcd_snap_fsync_duration_seconds_bucket{le=\"0.001\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.002\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.004\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.008\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.016\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.032\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.064\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.128\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.256\"} etcd_snap_fsync_duration_seconds_bucket{le=\"0.512\"} etcd_snap_fsync_duration_seconds_bucket{le=\"1.024\"} etcd_snap_fsync_duration_seconds_bucket{le=\"2.048\"} etcd_snap_fsync_duration_seconds_bucket{le=\"4.096\"} etcd_snap_fsync_duration_seconds_bucket{le=\"8.192\"} etcd_snap_fsync_duration_seconds_bucket{le=\"+Inf\"} etcd_snap_fsync_duration_seconds_sum etcd_snap_fsync_duration_seconds_count # HELP go_gc_duration_seconds A summary of the GC invocation durations. # TYPE go_gc_duration_seconds summary go_gc_duration_seconds{quantile=\"0\"} 0.00015764 go_gc_duration_seconds{quantile=\"0.25\"} 0.00015764 go_gc_duration_seconds{quantile=\"0.5\"} 0.000177495 go_gc_duration_seconds{quantile=\"0.75\"} 0.000184464 go_gc_duration_seconds{quantile=\"1\"} 0.000184464 go_gc_duration_seconds_sum 0.000519599 go_gc_duration_seconds_count 3 # HELP go_goroutines Number of goroutines that currently exist. # TYPE go_goroutines gauge go_goroutines 47 # HELP go_info Information about the Go environment. # TYPE go_info gauge go_info{version=\"devel +b74f732 Mon Jun 11 00:45:20 2018 +0000\"} 1 # HELP go_memstats_alloc_bytes Number of bytes allocated and still in use. # TYPE go_memstats_alloc_bytes gauge go_memstats_alloc_bytes 4.150712e+06 # HELP go_memstats_alloc_bytes_total Total number of bytes allocated, even if freed. # TYPE go_memstats_alloc_bytes_total counter go_memstats_alloc_bytes_total 8.208744e+06 # HELP go_memstats_buck_hash_sys_bytes Number of bytes used by the profiling bucket hash table. # TYPE go_memstats_buck_hash_sys_bytes gauge go_memstats_buck_hash_sys_bytes 1.446981e+06 # HELP go_memstats_frees_total Total number of frees. # TYPE go_memstats_frees_total counter go_memstats_frees_total 996 # HELP go_memstats_gc_cpu_fraction The fraction of this program's available CPU time used by the GC since the program started. # TYPE go_memstats_gc_cpu_fraction gauge go_memstats_gc_cpu_fraction 0.001566909347503071 # HELP go_memstats_gc_sys_bytes Number of bytes used for garbage collection system metadata. # TYPE go_memstats_gc_sys_bytes gauge go_memstats_gc_sys_bytes 2.371584e+06 # HELP go_memstats_heap_alloc_bytes Number of heap bytes allocated and still in use. # TYPE go_memstats_heap_alloc_bytes gauge go_memstats_heap_alloc_bytes 4.150712e+06 # HELP go_memstats_heap_idle_bytes Number of heap bytes waiting to be used. # TYPE go_memstats_heap_idle_bytes gauge go_memstats_heap_idle_bytes 5.9342848e+07 # HELP go_memstats_heap_inuse_bytes Number of heap bytes that are in use. # TYPE go_memstats_heap_inuse_bytes gauge go_memstats_heap_inuse_bytes 6.22592e+06 # HELP go_memstats_heap_objects Number of allocated objects. # TYPE go_memstats_heap_objects gauge go_memstats_heap_objects 2812 # HELP go_memstats_heap_released_bytes Number of heap bytes released to OS. # TYPE go_memstats_heap_released_bytes gauge go_memstats_heap_released_bytes # HELP go_memstats_heap_sys_bytes Number of heap bytes obtained from system. # TYPE go_memstats_heap_sys_bytes gauge go_memstats_heap_sys_bytes 6.5568768e+07 # HELP go_memstats_last_gc_time_seconds Number of seconds since 1970 of last garbage collection. # TYPE go_memstats_last_gc_time_seconds gauge go_memstats_last_gc_time_seconds 1.5286935487193735e+09 # HELP go_memstats_lookups_total Total number of pointer lookups. # TYPE go_memstats_lookups_total counter go_memstats_lookups_total # HELP go_memstats_mallocs_total Total number of mallocs. # TYPE go_memstats_mallocs_total counter go_memstats_mallocs_total 3808 # HELP go_memstats_mcache_inuse_bytes Number of bytes in use by mcache structures. # TYPE go_memstats_mcache_inuse_bytes gauge go_memstats_mcache_inuse_bytes 27648 # HELP go_memstats_mcache_sys_bytes Number of bytes used for mcache structures obtained from system. # TYPE go_memstats_mcache_sys_bytes gauge go_memstats_mcache_sys_bytes 32768 # HELP go_memstats_mspan_inuse_bytes Number of bytes in use by mspan structures. # TYPE go_memstats_mspan_inuse_bytes gauge go_memstats_mspan_inuse_bytes 74784 # HELP go_memstats_mspan_sys_bytes Number of bytes used for mspan structures obtained from system. # TYPE go_memstats_mspan_sys_bytes gauge go_memstats_mspan_sys_bytes 8192 # HELP go_memstats_next_gc_bytes Number of heap bytes when next garbage collection will take place. # TYPE go_memstats_next_gc_bytes gauge go_memstats_next_gc_bytes 6.983584e+06 # HELP go_memstats_other_sys_bytes Number of bytes used for other system allocations. # TYPE go_memstats_other_sys_bytes gauge go_memstats_other_sys_bytes 2.290867e+06 # HELP go_memstats_stack_inuse_bytes Number of bytes in use by the stack allocator. # TYPE go_memstats_stack_inuse_bytes gauge go_memstats_stack_inuse_bytes 1.540096e+06 # HELP go_memstats_stack_sys_bytes Number of bytes obtained from system for stack allocator. # TYPE go_memstats_stack_sys_bytes gauge go_memstats_stack_sys_bytes 1.540096e+06 # HELP go_memstats_sys_bytes Number of bytes obtained from system. # TYPE go_memstats_sys_bytes gauge go_memstats_sys_bytes 7.3332984e+07 # HELP go_threads Number of OS threads created. # TYPE go_threads gauge go_threads 23 # HELP grpc_server_handled_total Total number of RPCs completed on the server, regardless of success or failure. # TYPE grpc_server_handled_total counter grpc_server_handled_total{grpc_code=\"Aborted\",grpc_method=\"Alarm\",grpc_service=\"etcdserverpb.Maintenance\",grpc_type=\"unary\"} grpc_server_handled_total{grpc_code=\"Aborted\",grpc_method=\"AuthDisable\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} <snip> grpc_server_started_total{grpc_method=\"UserList\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} grpc_server_started_total{grpc_method=\"UserRevokeRole\",grpc_service=\"etcdserverpb.Auth\",grpc_type=\"unary\"} grpc_server_started_total{grpc_method=\"Watch\",grpc_service=\"etcdserverpb.Watch\",grpc_type=\"bidi_stream\"} # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. # TYPE process_cpu_seconds_total counter process_cpu_seconds_total 0.12 # HELP process_max_fds Maximum number of open file descriptors. # TYPE process_max_fds gauge process_max_fds 1.048576e+06 # HELP process_open_fds Number of open file descriptors. # TYPE process_open_fds gauge process_open_fds 14 # HELP process_resident_memory_bytes Resident memory size in bytes. # TYPE process_resident_memory_bytes gauge process_resident_memory_bytes 1.7301504e+07 # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. # TYPE process_start_time_seconds gauge process_start_time_seconds 1.52869354775e+09 # HELP process_virtual_memory_bytes Virtual memory size in bytes. # TYPE process_virtual_memory_bytes gauge process_virtual_memory_bytes 1.08675072e+1 "])) === RUN TestCtlV2Set sudeeshjohn:/tmp$ ``` etcd version: 3.3.0+git
test
etcd integration test is failing test log pass run fail curl test go failed get with curl read dev ptmx input output error expected alert handshake failure got help etcd debugging lease granted total the total number of granted leases type etcd debugging lease granted total counter etcd debugging lease granted total help etcd debugging lease renewed total the number of renewed leases seen by the leader type etcd debugging lease renewed total counter etcd debugging lease renewed total help etcd debugging lease revoked total the total number of revoked leases type etcd debugging lease revoked total counter etcd debugging lease revoked total help etcd debugging lease ttl total bucketed histogram of lease ttls type etcd debugging lease ttl total histogram etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le etcd debugging lease ttl total bucket le inf etcd debugging lease ttl total sum etcd debugging lease ttl total count help etcd debugging mvcc db compaction keys total total number of db keys compacted type etcd debugging mvcc db compaction keys total counter etcd debugging mvcc db compaction keys total help etcd debugging mvcc db compaction pause duration milliseconds bucketed histogram of db compaction pause duration type etcd debugging mvcc db compaction pause duration milliseconds histogram etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le etcd debugging mvcc db compaction pause duration milliseconds bucket le inf etcd debugging mvcc db compaction pause duration milliseconds sum etcd debugging mvcc db compaction pause duration milliseconds count help etcd debugging mvcc db compaction total duration milliseconds bucketed histogram of db compaction total duration type etcd debugging mvcc db compaction total duration milliseconds histogram etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le etcd debugging mvcc db compaction total duration milliseconds bucket le inf etcd debugging mvcc db compaction total duration milliseconds sum etcd debugging mvcc db compaction total duration milliseconds count help etcd debugging mvcc db total size in bytes total size of the underlying database physically allocated in bytes type etcd debugging mvcc db total size in bytes gauge etcd debugging mvcc db total size in bytes help etcd debugging mvcc delete total total number of deletes seen by this member type etcd debugging mvcc delete total counter etcd debugging mvcc delete total help etcd debugging mvcc events total total number of events sent by this member type etcd debugging mvcc events total counter etcd debugging mvcc events total help etcd debugging mvcc index compaction pause duration milliseconds bucketed histogram of index compaction pause duration type etcd debugging mvcc index compaction pause duration milliseconds histogram etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le etcd debugging mvcc index compaction pause duration milliseconds bucket le inf etcd debugging mvcc index compaction pause duration milliseconds sum etcd debugging mvcc index compaction pause duration milliseconds count help etcd debugging mvcc keys total total number of keys type etcd debugging mvcc keys total gauge etcd debugging mvcc keys total help etcd debugging mvcc pending events total total number of pending events to be sent type etcd debugging mvcc pending events total gauge etcd debugging mvcc pending events total help etcd debugging mvcc put total total number of puts seen by this member type etcd debugging mvcc put total counter etcd debugging mvcc put total help etcd debugging mvcc range total total number of ranges seen by this member type etcd debugging mvcc range total counter etcd debugging mvcc range total help etcd debugging mvcc slow watcher total total number of unsynced slow watchers type etcd debugging mvcc slow watcher total gauge etcd debugging mvcc slow watcher total help etcd debugging mvcc txn total total number of txns seen by this member type etcd debugging mvcc txn total counter etcd debugging mvcc txn total help etcd debugging mvcc watch stream total total number of watch streams type etcd debugging mvcc watch stream total gauge etcd debugging mvcc watch stream total help etcd debugging mvcc watcher total total number of watchers type etcd debugging mvcc watcher total gauge etcd debugging mvcc watcher total help etcd debugging server lease expired total the total number of expired leases type etcd debugging server lease expired total counter etcd debugging server lease expired total help etcd debugging snap save marshalling duration seconds the marshalling cost distributions of save called by snapshot type etcd debugging snap save marshalling duration seconds histogram etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le etcd debugging snap save marshalling duration seconds bucket le inf etcd debugging snap save marshalling duration seconds sum etcd debugging snap save marshalling duration seconds count help etcd debugging snap save total duration seconds the total latency distributions of save called by snapshot type etcd debugging snap save total duration seconds histogram etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le etcd debugging snap save total duration seconds bucket le inf etcd debugging snap save total duration seconds sum etcd debugging snap save total duration seconds count help etcd debugging store expires total total number of expired keys type etcd debugging store expires total counter etcd debugging store expires total help etcd debugging store reads total total number of reads action by get getrecursive local to this member type etcd debugging store reads total counter etcd debugging store reads total action getrecursive help etcd debugging store watch requests total total number of incoming watch requests new or reestablished type etcd debugging store watch requests total counter etcd debugging store watch requests total help etcd debugging store watchers count of currently active watchers type etcd debugging store watchers gauge etcd debugging store watchers help etcd debugging store writes total total number of writes e g set compareanddelete seen by this member type etcd debugging store writes total counter etcd debugging store writes total action create etcd debugging store writes total action set help etcd disk backend commit duration seconds the latency distributions of commit called by backend type etcd disk backend commit duration seconds histogram etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le etcd disk backend commit duration seconds bucket le inf etcd disk backend commit duration seconds sum etcd disk backend commit duration seconds count help etcd disk backend defrag duration seconds the latency distribution of backend defragmentation type etcd disk backend defrag duration seconds histogram etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le etcd disk backend defrag duration seconds bucket le inf etcd disk backend defrag duration seconds sum etcd disk backend defrag duration seconds count help etcd disk backend snapshot duration seconds the latency distribution of backend snapshots type etcd disk backend snapshot duration seconds histogram etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le etcd disk backend snapshot duration seconds bucket le inf etcd disk backend snapshot duration seconds sum etcd disk backend snapshot duration seconds count help etcd disk wal fsync duration seconds the latency distributions of fsync called by wal type etcd disk wal fsync duration seconds histogram etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le etcd disk wal fsync duration seconds bucket le inf etcd disk wal fsync duration seconds sum etcd disk wal fsync duration seconds count help etcd grpc proxy cache hits total total number of cache hits type etcd grpc proxy cache hits total gauge etcd grpc proxy cache hits total help etcd grpc proxy cache keys total total number of keys ranges cached type etcd grpc proxy cache keys total gauge etcd grpc proxy cache keys total help etcd grpc proxy cache misses total total number of cache misses type etcd grpc proxy cache misses total gauge etcd grpc proxy cache misses total help etcd grpc proxy events coalescing total total number of events coalescing type etcd grpc proxy events coalescing total counter etcd grpc proxy events coalescing total help etcd grpc proxy watchers coalescing total total number of current watchers coalescing type etcd grpc proxy watchers coalescing total gauge etcd grpc proxy watchers coalescing total help etcd mvcc db total size in bytes total size of the underlying database physically allocated in bytes type etcd mvcc db total size in bytes gauge etcd mvcc db total size in bytes help etcd mvcc db total size in use in bytes total size of the underlying database logically in use in bytes type etcd mvcc db total size in use in bytes gauge etcd mvcc db total size in use in bytes help etcd mvcc hash duration seconds the latency distribution of storage hash operation type etcd mvcc hash duration seconds histogram etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le etcd mvcc hash duration seconds bucket le inf etcd mvcc hash duration seconds sum etcd mvcc hash duration seconds count help etcd mvcc hash rev duration seconds the latency distribution of storage hash by revision operation type etcd mvcc hash rev duration seconds histogram etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le etcd mvcc hash rev duration seconds bucket le inf etcd mvcc hash rev duration seconds sum etcd mvcc hash rev duration seconds count help etcd network client grpc received bytes total the total number of bytes received from grpc clients type etcd network client grpc received bytes total counter etcd network client grpc received bytes total help etcd network client grpc sent bytes total the total number of bytes sent to grpc clients type etcd network client grpc sent bytes total counter etcd network client grpc sent bytes total help etcd server has leader whether or not a leader exists is existence is not type etcd server has leader gauge etcd server has leader help etcd server heartbeat send failures total the total number of leader heartbeat send failures likely overloaded from slow disk type etcd server heartbeat send failures total counter etcd server heartbeat send failures total help etcd server is leader whether or not this member is a leader if is otherwise type etcd server is leader gauge etcd server is leader help etcd server leader changes seen total the number of leader changes seen type etcd server leader changes seen total counter etcd server leader changes seen total help etcd server proposals applied total the total number of consensus proposals applied type etcd server proposals applied total gauge etcd server proposals applied total help etcd server proposals committed total the total number of consensus proposals committed type etcd server proposals committed total gauge etcd server proposals committed total help etcd server proposals failed total the total number of failed proposals seen type etcd server proposals failed total counter etcd server proposals failed total help etcd server proposals pending the current number of pending proposals to commit type etcd server proposals pending gauge etcd server proposals pending help etcd server quota backend bytes current backend storage quota size in bytes type etcd server quota backend bytes gauge etcd server quota backend bytes help etcd server slow apply total the total number of slow apply requests likely overloaded from slow disk type etcd server slow apply total counter etcd server slow apply total help etcd server version which version is running for server version label with current version type etcd server version gauge etcd server version server version git help etcd snap fsync duration seconds the latency distributions of fsync called by snap type etcd snap fsync duration seconds histogram etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le etcd snap fsync duration seconds bucket le inf etcd snap fsync duration seconds sum etcd snap fsync duration seconds count help go gc duration seconds a summary of the gc invocation durations type go gc duration seconds summary go gc duration seconds quantile go gc duration seconds quantile go gc duration seconds quantile go gc duration seconds quantile go gc duration seconds quantile go gc duration seconds sum go gc duration seconds count help go goroutines number of goroutines that currently exist type go goroutines gauge go goroutines help go info information about the go environment type go info gauge go info version devel mon jun help go memstats alloc bytes number of bytes allocated and still in use type go memstats alloc bytes gauge go memstats alloc bytes help go memstats alloc bytes total total number of bytes allocated even if freed type go memstats alloc bytes total counter go memstats alloc bytes total help go memstats buck hash sys bytes number of bytes used by the profiling bucket hash table type go memstats buck hash sys bytes gauge go memstats buck hash sys bytes help go memstats frees total total number of frees type go memstats frees total counter go memstats frees total help go memstats gc cpu fraction the fraction of this program s available cpu time used by the gc since the program started type go memstats gc cpu fraction gauge go memstats gc cpu fraction help go memstats gc sys bytes number of bytes used for garbage collection system metadata type go memstats gc sys bytes gauge go memstats gc sys bytes help go memstats heap alloc bytes number of heap bytes allocated and still in use type go memstats heap alloc bytes gauge go memstats heap alloc bytes help go memstats heap idle bytes number of heap bytes waiting to be used type go memstats heap idle bytes gauge go memstats heap idle bytes help go memstats heap inuse bytes number of heap bytes that are in use type go memstats heap inuse bytes gauge go memstats heap inuse bytes help go memstats heap objects number of allocated objects type go memstats heap objects gauge go memstats heap objects help go memstats heap released bytes number of heap bytes released to os type go memstats heap released bytes gauge go memstats heap released bytes help go memstats heap sys bytes number of heap bytes obtained from system type go memstats heap sys bytes gauge go memstats heap sys bytes help go memstats last gc time seconds number of seconds since of last garbage collection type go memstats last gc time seconds gauge go memstats last gc time seconds help go memstats lookups total total number of pointer lookups type go memstats lookups total counter go memstats lookups total help go memstats mallocs total total number of mallocs type go memstats mallocs total counter go memstats mallocs total help go memstats mcache inuse bytes number of bytes in use by mcache structures type go memstats mcache inuse bytes gauge go memstats mcache inuse bytes help go memstats mcache sys bytes number of bytes used for mcache structures obtained from system type go memstats mcache sys bytes gauge go memstats mcache sys bytes help go memstats mspan inuse bytes number of bytes in use by mspan structures type go memstats mspan inuse bytes gauge go memstats mspan inuse bytes help go memstats mspan sys bytes number of bytes used for mspan structures obtained from system type go memstats mspan sys bytes gauge go memstats mspan sys bytes help go memstats next gc bytes number of heap bytes when next garbage collection will take place type go memstats next gc bytes gauge go memstats next gc bytes help go memstats other sys bytes number of bytes used for other system allocations type go memstats other sys bytes gauge go memstats other sys bytes help go memstats stack inuse bytes number of bytes in use by the stack allocator type go memstats stack inuse bytes gauge go memstats stack inuse bytes help go memstats stack sys bytes number of bytes obtained from system for stack allocator type go memstats stack sys bytes gauge go memstats stack sys bytes help go memstats sys bytes number of bytes obtained from system type go memstats sys bytes gauge go memstats sys bytes help go threads number of os threads created type go threads gauge go threads help grpc server handled total total number of rpcs completed on the server regardless of success or failure type grpc server handled total counter grpc server handled total grpc code aborted grpc method alarm grpc service etcdserverpb maintenance grpc type unary grpc server handled total grpc code aborted grpc method authdisable grpc service etcdserverpb auth grpc type unary grpc server started total grpc method userlist grpc service etcdserverpb auth grpc type unary grpc server started total grpc method userrevokerole grpc service etcdserverpb auth grpc type unary grpc server started total grpc method watch grpc service etcdserverpb watch grpc type bidi stream help process cpu seconds total total user and system cpu time spent in seconds type process cpu seconds total counter process cpu seconds total help process max fds maximum number of open file descriptors type process max fds gauge process max fds help process open fds number of open file descriptors type process open fds gauge process open fds help process resident memory bytes resident memory size in bytes type process resident memory bytes gauge process resident memory bytes help process start time seconds start time of the process since unix epoch in seconds type process start time seconds gauge process start time seconds help process virtual memory bytes virtual memory size in bytes type process virtual memory bytes gauge process virtual memory bytes run sudeeshjohn tmp etcd version git
1
159,176
20,036,683,502
IssuesEvent
2022-02-02 12:40:35
kapseliboi/broker
https://api.github.com/repos/kapseliboi/broker
opened
CVE-2020-7788 (High) detected in ini-1.3.5.tgz
security vulnerability
## CVE-2020-7788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ini-1.3.5.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-datachannel/node_modules/ini/package.json</p> <p> Dependency Hierarchy: - streamr-network-24.3.0.tgz (Root Library) - node-datachannel-0.1.4.tgz - prebuild-install-5.3.6.tgz - rc-1.2.8.tgz - :x: **ini-1.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/broker/commit/f1837aa3159e17151d81e973e4c79872a4bdf782">f1837aa3159e17151d81e973e4c79872a4bdf782</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: v1.3.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7788 (High) detected in ini-1.3.5.tgz - ## CVE-2020-7788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ini-1.3.5.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-datachannel/node_modules/ini/package.json</p> <p> Dependency Hierarchy: - streamr-network-24.3.0.tgz (Root Library) - node-datachannel-0.1.4.tgz - prebuild-install-5.3.6.tgz - rc-1.2.8.tgz - :x: **ini-1.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/broker/commit/f1837aa3159e17151d81e973e4c79872a4bdf782">f1837aa3159e17151d81e973e4c79872a4bdf782</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: v1.3.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ini tgz cve high severity vulnerability vulnerable library ini tgz an ini encoder decoder for node library home page a href path to dependency file package json path to vulnerable library node modules node datachannel node modules ini package json dependency hierarchy streamr network tgz root library node datachannel tgz prebuild install tgz rc tgz x ini tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package ini before if an attacker submits a malicious ini file to an application that parses it with ini parse they will pollute the prototype on the application this can be exploited further depending on the context publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
148,859
11,868,556,678
IssuesEvent
2020-03-26 09:24:42
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] createNodeKeyStore failed on Jenkins darwin worker due to DNS name
:Query Languages/SQL >test-failure
https://gradle-enterprise.elastic.co/s/r3hcjn6yivexe failed with the following error in a darwin build (https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-darwin-compatibility/595): ``` 00:08:37 keytool error: java.lang.RuntimeException: java.io.IOException: DNSName components must begin with a letter 00:08:37 00:08:37 build complete, generating: /private/var/lib/jenkins/workspace/elastic+elasticsearch+master+multijob-darwin-compatibility/build/595.tar.bz2 00:09:40 00:09:40 FAILURE: Build failed with an exception. 00:09:40 00:09:40 * What went wrong: 00:09:40 Execution failed for task ':x-pack:plugin:sql:qa:security:with-ssl:createNodeKeyStore'. 00:09:40 > Process '/var/lib/jenkins/.java/openjdk11/bin/keytool [-genkey, -alias, test-node, -keystore, /private/var/lib/jenkins/workspace/elastic+elasticsearch+master+multijob-darwin-compatibility/x-pack/plugin/sql/qa/security/with-ssl/build/keystore/test-node.jks, -keyalg, RSA, -keysize, 2048, -validity, 712, -dname, CN=smoke-test-plugins-ssl, -keypass, keypass, -storepass, keypass, -ext, san=ip:fe80::1,dns:20049.local,ip:::1,dns:localhost,ip:127.0.0.1,dns:localhost]' finished with non-zero exit value 1 ``` I am guessing the part of that that `keytool` took offence at is `dns:20049.local`. I can run `./gradlew ':x-pack:plugin:sql:qa:security:with-ssl:createNodeKeyStore'` without problems on my Macbook. Presumably there is something different about the way the macOS Jenkins workers are configured that means they have a technically invalid DNS name.
1.0
[CI] createNodeKeyStore failed on Jenkins darwin worker due to DNS name - https://gradle-enterprise.elastic.co/s/r3hcjn6yivexe failed with the following error in a darwin build (https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-darwin-compatibility/595): ``` 00:08:37 keytool error: java.lang.RuntimeException: java.io.IOException: DNSName components must begin with a letter 00:08:37 00:08:37 build complete, generating: /private/var/lib/jenkins/workspace/elastic+elasticsearch+master+multijob-darwin-compatibility/build/595.tar.bz2 00:09:40 00:09:40 FAILURE: Build failed with an exception. 00:09:40 00:09:40 * What went wrong: 00:09:40 Execution failed for task ':x-pack:plugin:sql:qa:security:with-ssl:createNodeKeyStore'. 00:09:40 > Process '/var/lib/jenkins/.java/openjdk11/bin/keytool [-genkey, -alias, test-node, -keystore, /private/var/lib/jenkins/workspace/elastic+elasticsearch+master+multijob-darwin-compatibility/x-pack/plugin/sql/qa/security/with-ssl/build/keystore/test-node.jks, -keyalg, RSA, -keysize, 2048, -validity, 712, -dname, CN=smoke-test-plugins-ssl, -keypass, keypass, -storepass, keypass, -ext, san=ip:fe80::1,dns:20049.local,ip:::1,dns:localhost,ip:127.0.0.1,dns:localhost]' finished with non-zero exit value 1 ``` I am guessing the part of that that `keytool` took offence at is `dns:20049.local`. I can run `./gradlew ':x-pack:plugin:sql:qa:security:with-ssl:createNodeKeyStore'` without problems on my Macbook. Presumably there is something different about the way the macOS Jenkins workers are configured that means they have a technically invalid DNS name.
test
createnodekeystore failed on jenkins darwin worker due to dns name failed with the following error in a darwin build keytool error java lang runtimeexception java io ioexception dnsname components must begin with a letter build complete generating private var lib jenkins workspace elastic elasticsearch master multijob darwin compatibility build tar failure build failed with an exception what went wrong execution failed for task x pack plugin sql qa security with ssl createnodekeystore process var lib jenkins java bin keytool finished with non zero exit value i am guessing the part of that that keytool took offence at is dns local i can run gradlew x pack plugin sql qa security with ssl createnodekeystore without problems on my macbook presumably there is something different about the way the macos jenkins workers are configured that means they have a technically invalid dns name
1
55,434
6,475,980,459
IssuesEvent
2017-08-17 21:39:15
RocketChat/Rocket.Chat.Electron
https://api.github.com/repos/RocketChat/Rocket.Chat.Electron
closed
Uncaught exception when opening "About Rocket.Chat+"
plat: linux stat: needs testing
## My Setup - Operating System: Ubuntu 17.04 - App Version: 2.7.0 - [x] I have tested with the latest version - [x] I can simulate the issue easily ## Description When opening the About Rocket.Chat+ menu option the "About" window opens, empty, and I also get an error window "A JavaScript error occurred in the main process" with the following content: ```Uncaught Exception: TypeError: Error processing argument at index 0, conversion failure from undefined at Event.set (/opt/Rocket.Chat+/resources/electron.asar/browser/api/web-contents.js:252:22) at EventEmitter.electron.ipcMain.on (/opt/Rocket.Chat+/resources/app.asar/app/background.js:496:27) at emitTwo (events.js:106:13) at EventEmitter.emit (events.js:191:7) at WebContents.<anonymous> (/opt/Rocket.Chat+/resources/electron.asar/browser/api/web-contents.js:256:37) at emitTwo (events.js:106:13) at WebContents.emit (events.js:191:7) ``` I believe this did not happen with previous versions, but started with 2.7.0.
1.0
Uncaught exception when opening "About Rocket.Chat+" - ## My Setup - Operating System: Ubuntu 17.04 - App Version: 2.7.0 - [x] I have tested with the latest version - [x] I can simulate the issue easily ## Description When opening the About Rocket.Chat+ menu option the "About" window opens, empty, and I also get an error window "A JavaScript error occurred in the main process" with the following content: ```Uncaught Exception: TypeError: Error processing argument at index 0, conversion failure from undefined at Event.set (/opt/Rocket.Chat+/resources/electron.asar/browser/api/web-contents.js:252:22) at EventEmitter.electron.ipcMain.on (/opt/Rocket.Chat+/resources/app.asar/app/background.js:496:27) at emitTwo (events.js:106:13) at EventEmitter.emit (events.js:191:7) at WebContents.<anonymous> (/opt/Rocket.Chat+/resources/electron.asar/browser/api/web-contents.js:256:37) at emitTwo (events.js:106:13) at WebContents.emit (events.js:191:7) ``` I believe this did not happen with previous versions, but started with 2.7.0.
test
uncaught exception when opening about rocket chat my setup operating system ubuntu app version i have tested with the latest version i can simulate the issue easily description when opening the about rocket chat menu option the about window opens empty and i also get an error window a javascript error occurred in the main process with the following content uncaught exception typeerror error processing argument at index conversion failure from undefined at event set opt rocket chat resources electron asar browser api web contents js at eventemitter electron ipcmain on opt rocket chat resources app asar app background js at emittwo events js at eventemitter emit events js at webcontents opt rocket chat resources electron asar browser api web contents js at emittwo events js at webcontents emit events js i believe this did not happen with previous versions but started with
1
304,440
26,276,529,389
IssuesEvent
2023-01-06 22:48:23
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
compose/compare/compare: TestCompare failed
C-test-failure O-robot T-sql-sessions branch-release-22.2.0
compose/compare/compare.TestCompare [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7800161?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7800161?buildTab=artifacts#/) on release-22.2.0 @ [413dc596348780a8c81a450dc135135985ed70af](https://github.com/cockroachdb/cockroach/commits/413dc596348780a8c81a450dc135135985ed70af): ``` === RUN TestCompare compare_test.go:119: Checking connection to: cockroach1 I221202 07:29:31.305388 6644 testutils/soon.go:40 [-] 1 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:32.311448 6644 testutils/soon.go:40 [-] 2 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:33.318711 6644 testutils/soon.go:40 [-] 3 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:34.328892 6644 testutils/soon.go:40 [-] 4 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:35.340001 6644 testutils/soon.go:40 [-] 5 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:36.348195 6644 testutils/soon.go:40 [-] 6 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:37.355682 6644 testutils/soon.go:40 [-] 7 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:38.364528 6644 testutils/soon.go:40 [-] 8 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:39.372193 6644 testutils/soon.go:40 [-] 9 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:40.377310 6644 testutils/soon.go:40 [-] 10 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:41.389140 6644 testutils/soon.go:40 [-] 11 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:42.394586 6644 testutils/soon.go:40 [-] 12 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:43.401772 6644 testutils/soon.go:40 [-] 13 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:44.408743 6644 testutils/soon.go:40 [-] 14 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:45.417169 6644 testutils/soon.go:40 [-] 15 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:46.424574 6644 testutils/soon.go:40 [-] 16 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:47.430855 6644 testutils/soon.go:40 [-] 17 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:48.437342 6644 testutils/soon.go:40 [-] 18 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:49.444158 6644 testutils/soon.go:40 [-] 19 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:50.454701 6644 testutils/soon.go:40 [-] 20 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:51.461972 6644 testutils/soon.go:40 [-] 21 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:52.472098 6644 testutils/soon.go:40 [-] 22 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:53.479802 6644 testutils/soon.go:40 [-] 23 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:54.488111 6644 testutils/soon.go:40 [-] 24 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:55.494594 6644 testutils/soon.go:40 [-] 25 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:56.501172 6644 testutils/soon.go:40 [-] 26 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:57.506943 6644 testutils/soon.go:40 [-] 27 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:58.516928 6644 testutils/soon.go:40 [-] 28 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:59.522566 6644 testutils/soon.go:40 [-] 29 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:00.529021 6644 testutils/soon.go:40 [-] 30 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:01.535990 6644 testutils/soon.go:40 [-] 31 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:02.542401 6644 testutils/soon.go:40 [-] 32 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:03.549572 6644 testutils/soon.go:40 [-] 33 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:04.555510 6644 testutils/soon.go:40 [-] 34 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:05.565682 6644 testutils/soon.go:40 [-] 35 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:06.573276 6644 testutils/soon.go:40 [-] 36 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:07.586920 6644 testutils/soon.go:40 [-] 37 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:08.591984 6644 testutils/soon.go:40 [-] 38 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:09.600497 6644 testutils/soon.go:40 [-] 39 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:10.606837 6644 testutils/soon.go:40 [-] 40 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:11.613003 6644 testutils/soon.go:40 [-] 41 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:12.619454 6644 testutils/soon.go:40 [-] 42 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host compare_test.go:120: condition failed to evaluate within 45s: failed to connect to ``host=cockroach1 user=root database=postgres``: hostname resolving error (lookup cockroach1 on 169.254.169.254:53: no such host) --- FAIL: TestCompare (45.60s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #92815 compose/compare/compare: TestCompare failed [C-test-failure O-robot T-sql-experience branch-release-22.2] - #92044 compose/compare/compare: TestCompare failed [C-test-failure O-robot T-sql-experience branch-master] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestCompare.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22028
1.0
compose/compare/compare: TestCompare failed - compose/compare/compare.TestCompare [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7800161?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7800161?buildTab=artifacts#/) on release-22.2.0 @ [413dc596348780a8c81a450dc135135985ed70af](https://github.com/cockroachdb/cockroach/commits/413dc596348780a8c81a450dc135135985ed70af): ``` === RUN TestCompare compare_test.go:119: Checking connection to: cockroach1 I221202 07:29:31.305388 6644 testutils/soon.go:40 [-] 1 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:32.311448 6644 testutils/soon.go:40 [-] 2 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:33.318711 6644 testutils/soon.go:40 [-] 3 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:34.328892 6644 testutils/soon.go:40 [-] 4 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:35.340001 6644 testutils/soon.go:40 [-] 5 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:36.348195 6644 testutils/soon.go:40 [-] 6 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:37.355682 6644 testutils/soon.go:40 [-] 7 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:38.364528 6644 testutils/soon.go:40 [-] 8 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:39.372193 6644 testutils/soon.go:40 [-] 9 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:40.377310 6644 testutils/soon.go:40 [-] 10 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:41.389140 6644 testutils/soon.go:40 [-] 11 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:42.394586 6644 testutils/soon.go:40 [-] 12 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:43.401772 6644 testutils/soon.go:40 [-] 13 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:44.408743 6644 testutils/soon.go:40 [-] 14 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:45.417169 6644 testutils/soon.go:40 [-] 15 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:46.424574 6644 testutils/soon.go:40 [-] 16 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:47.430855 6644 testutils/soon.go:40 [-] 17 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:48.437342 6644 testutils/soon.go:40 [-] 18 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:49.444158 6644 testutils/soon.go:40 [-] 19 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:50.454701 6644 testutils/soon.go:40 [-] 20 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:51.461972 6644 testutils/soon.go:40 [-] 21 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:52.472098 6644 testutils/soon.go:40 [-] 22 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:53.479802 6644 testutils/soon.go:40 [-] 23 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:54.488111 6644 testutils/soon.go:40 [-] 24 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:55.494594 6644 testutils/soon.go:40 [-] 25 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:56.501172 6644 testutils/soon.go:40 [-] 26 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:57.506943 6644 testutils/soon.go:40 [-] 27 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:58.516928 6644 testutils/soon.go:40 [-] 28 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:29:59.522566 6644 testutils/soon.go:40 [-] 29 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:00.529021 6644 testutils/soon.go:40 [-] 30 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:01.535990 6644 testutils/soon.go:40 [-] 31 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:02.542401 6644 testutils/soon.go:40 [-] 32 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:03.549572 6644 testutils/soon.go:40 [-] 33 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:04.555510 6644 testutils/soon.go:40 [-] 34 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:05.565682 6644 testutils/soon.go:40 [-] 35 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:06.573276 6644 testutils/soon.go:40 [-] 36 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:07.586920 6644 testutils/soon.go:40 [-] 37 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:08.591984 6644 testutils/soon.go:40 [-] 38 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:09.600497 6644 testutils/soon.go:40 [-] 39 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:10.606837 6644 testutils/soon.go:40 [-] 40 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:11.613003 6644 testutils/soon.go:40 [-] 41 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host I221202 07:30:12.619454 6644 testutils/soon.go:40 [-] 42 SucceedsSoon: lookup cockroach1 on 169.254.169.254:53: no such host compare_test.go:120: condition failed to evaluate within 45s: failed to connect to ``host=cockroach1 user=root database=postgres``: hostname resolving error (lookup cockroach1 on 169.254.169.254:53: no such host) --- FAIL: TestCompare (45.60s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #92815 compose/compare/compare: TestCompare failed [C-test-failure O-robot T-sql-experience branch-release-22.2] - #92044 compose/compare/compare: TestCompare failed [C-test-failure O-robot T-sql-experience branch-master] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestCompare.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22028
test
compose compare compare testcompare failed compose compare compare testcompare with on release run testcompare compare test go checking connection to testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host testutils soon go succeedssoon lookup on no such host compare test go condition failed to evaluate within failed to connect to host user root database postgres hostname resolving error lookup on no such host fail testcompare parameters tags bazel gss deadlock help see also same failure on other branches compose compare compare testcompare failed compose compare compare testcompare failed cc cockroachdb sql experience jira issue crdb
1
261,756
22,773,001,706
IssuesEvent
2022-07-08 11:54:54
admiral-team/admiralui-android
https://api.github.com/repos/admiral-team/admiralui-android
closed
[FEATURE] - Badges screen clicker test
enhancement test
**Описание задачи:** Сделать тест с прокликиванием экрана Badges **Фреймворк** XML **Ресурсы:** Ссылка на макеты или ресурсы...
1.0
[FEATURE] - Badges screen clicker test - **Описание задачи:** Сделать тест с прокликиванием экрана Badges **Фреймворк** XML **Ресурсы:** Ссылка на макеты или ресурсы...
test
badges screen clicker test описание задачи сделать тест с прокликиванием экрана badges фреймворк xml ресурсы ссылка на макеты или ресурсы
1
224,955
17,785,503,772
IssuesEvent
2021-08-31 10:31:24
SAPDocuments/Issues
https://api.github.com/repos/SAPDocuments/Issues
closed
Req. Product Details - API portal provisioning issue - AP region
High-Prio SCPTest-trial3 SCPTest-cloudin SCPTest-2108A
Hello, Tutorial Link: https://developers.sap.com/tutorials/cp-starter-isuite-onboard-subscribe.html Step 5: Verify the activated capabilities Issue: After both the capabilities are active and the boosters are executed, we should be able to access them both. But I can't access API Portal. It is running into the error and says there is a temporary problem provisioning your account. (In AP region) Screenshot: ![image](https://user-images.githubusercontent.com/66831161/128850923-ba0392de-e019-4089-a0cb-913818e394a6.png) Please check the issue. Thanks, Shubham Labels to be added: Sprint Label: SCPTest-2108A Scenario Label: SCPTest-cloudin Landscape Label: AP Prio: High-prio
3.0
Req. Product Details - API portal provisioning issue - AP region - Hello, Tutorial Link: https://developers.sap.com/tutorials/cp-starter-isuite-onboard-subscribe.html Step 5: Verify the activated capabilities Issue: After both the capabilities are active and the boosters are executed, we should be able to access them both. But I can't access API Portal. It is running into the error and says there is a temporary problem provisioning your account. (In AP region) Screenshot: ![image](https://user-images.githubusercontent.com/66831161/128850923-ba0392de-e019-4089-a0cb-913818e394a6.png) Please check the issue. Thanks, Shubham Labels to be added: Sprint Label: SCPTest-2108A Scenario Label: SCPTest-cloudin Landscape Label: AP Prio: High-prio
test
req product details api portal provisioning issue ap region hello tutorial link step verify the activated capabilities issue after both the capabilities are active and the boosters are executed we should be able to access them both but i can t access api portal it is running into the error and says there is a temporary problem provisioning your account in ap region screenshot please check the issue thanks shubham labels to be added sprint label scptest scenario label scptest cloudin landscape label ap prio high prio
1
195,574
15,531,014,131
IssuesEvent
2021-03-13 21:32:16
Sustaingineering/Pump-Firmware
https://api.github.com/repos/Sustaingineering/Pump-Firmware
closed
Document a 'Getting Started' Page
common documentation
If someone joins our team, we need a self contained page that help that person get up to speed.
1.0
Document a 'Getting Started' Page - If someone joins our team, we need a self contained page that help that person get up to speed.
non_test
document a getting started page if someone joins our team we need a self contained page that help that person get up to speed
0
106,467
11,489,438,244
IssuesEvent
2020-02-11 15:30:42
11ty/eleventy
https://api.github.com/repos/11ty/eleventy
closed
Document example of Next/Previous blog post links (without pagination)
documentation needs-votes
#211 solved the Pagination Next/Previous docs example but we still need one without pagination.
1.0
Document example of Next/Previous blog post links (without pagination) - #211 solved the Pagination Next/Previous docs example but we still need one without pagination.
non_test
document example of next previous blog post links without pagination solved the pagination next previous docs example but we still need one without pagination
0
249,842
26,988,741,217
IssuesEvent
2023-02-09 18:04:31
MatBenfield/news
https://api.github.com/repos/MatBenfield/news
closed
[SecurityWeek] Application Security Protection for the Masses
SecurityWeek Stale
I’ve always found it entertaining that so many sales pitches are essentially a listing of features for the product or service being sold. The reason I find this entertaining is that for anyone who has worked on the customer side or has ever listened to customers, it is obvious that customers buy solutions, not products. Thus, the notion of showing off how proud you are of your product by rattling off a laundry list of features has always seemed a bit odd to me. In other words, customers have a number of different problems, issues, and challenges that they are looking to solve. They are not necessarily interested in all of the different things your product or service can do. Rather, they are interested in learning how your solution can help them address their strategic priorities and move forward on the goals they have set for their security and fraud problems. It is incumbent upon vendors to understand that and to make it easy for potential customers to understand that mapping. Along those lines, improving application security is a common goal customers have. As you might imagine, any solution geared towards improving the security of an application is going to be complex, consisting of many different moving parts. Thus, forcing customers to hunt for the components they need within your product data sheets and overviews is not going to be an effective way to convince those customers that you have a solution they might be in the market for. So what can vendors do to convince customers that they have a solution worth that customer’s time to evaluate? For starters, they can bundle various features into use cases that can be easily demonstrated to, evaluated, and consumed by customers. Along those lines, what would a bundle around the popular application security protection use case look like? While not an exhaustive list, here are some thoughts: - **App Proxy:** Putting a proxy in front of applications is perhaps one of the most basic application security requirements, and for good reason. Having an intermediary allows us to inspect and monitor traffic going to and from the application, as well as to block or filter as necessary for security purposes. - **Rate Limiting and Fast Access Control Lists (ACLs):** Flooding a site is an old standby of attackers. It is a primitive, yet effective tactic. Rate limiting is a relatively straightforward way to prevent this type of attack. Similarly, fast-performing Access Control Lists (ACLs) are another effective way to keep unwanted traffic at bay. - **Path Discovery:** Applying machine learning (ML) to traffic transiting the environment allows us to track the rate of requests, the identity of clients accessing applications, the size of the payloads being sent, and other important telemetry elements. Using ML allows us to identify and block nefarious traffic before it becomes a more serious issue – often in minutes as opposed to hours. - **Web Application Firewall:** WAF has become a required technology for application providers and should be included as a part of any application security bundle. - **L3/L4/L7 DDoS:** DDoS protection has also become a requirement for application providers and should also be included as part of any application security bundle. - **Bot Defense:** Advanced bots that know how to get around the defenses listed above can cause application providers monetary loss and reputation damage. As such, bot defense should also be included as part of an application security bundle. - **Auto-Certificates:** Speed of deploying applications is essential for remaining competitive, as is speed of protecting those applications. The ability to auto-issue certificates and to auto-register DNS for resources saves time, allowing application providers to go from no protection to full protection in a matter of minutes. - **Malicious User Detection:** Another great application for machine learning (ML) is quickly understanding which users and patterns appear to be behaving maliciously. This is something that often takes application providers hours or days to identify. With ML, this can be done in minutes, allowing those application providers to quickly take action and block/mitigate. - **Client-Side Defense:** Visibility into the end-user environment is something many application providers lack. The ability to inspect how JavaScript is being called, where requests are going, and what third party scripts are being called gives important insight that is extremely helpful for application security purposes. - **URI Routing:** The ability to quickly and easily control where certain requests are routing gives application providers the ability to block/control specific endpoints (URIs). No application security solution would be complete without this important feature. - **Service Policies:** Quick and easy policy deployment is a must for application security. The ability to chain together service policies as needed based on requirements, along with the ability to generate custom rules for steering traffic or allowing/denying traffic beyond the capabilities of the other defensive capabilities is another essential part of the total application security package. - **Synthetic Monitors:** How are applications performing externally? What are my customers experiencing? These are important questions that synthetic monitors allow a business to answer, which can quickly identify any issues that might affect the application. - **TLS Fingerprinting and Device Identification:** While IP addresses change frequently, TLS fingerprints and device identifiers change much more rarely. Thus, basing policies and rules on them rather than IP address makes a lot of sense when it comes to application security. - **Cross-Site Request Forgery Protection:** Scripts that operate cross-site can cause serious problems for application providers. Thus mitigating the risk they present should be part of any application security bundle as well. Securing applications is a top priority for nearly all businesses. While there are many routes to application security, bundles that allow security teams to quickly and easily secure applications and affect security posture in a self-service manner are becoming increasingly popular. These bundles inform application providers and allow them to make better, more informed decisions to improve security posture without introducing unnecessary friction to the end-user. The post [Application Security Protection for the Masses](https://www.securityweek.com/application-security-protection-for-the-masses/) appeared first on [SecurityWeek](https://www.securityweek.com). <https://www.securityweek.com/application-security-protection-for-the-masses/>
True
[SecurityWeek] Application Security Protection for the Masses - I’ve always found it entertaining that so many sales pitches are essentially a listing of features for the product or service being sold. The reason I find this entertaining is that for anyone who has worked on the customer side or has ever listened to customers, it is obvious that customers buy solutions, not products. Thus, the notion of showing off how proud you are of your product by rattling off a laundry list of features has always seemed a bit odd to me. In other words, customers have a number of different problems, issues, and challenges that they are looking to solve. They are not necessarily interested in all of the different things your product or service can do. Rather, they are interested in learning how your solution can help them address their strategic priorities and move forward on the goals they have set for their security and fraud problems. It is incumbent upon vendors to understand that and to make it easy for potential customers to understand that mapping. Along those lines, improving application security is a common goal customers have. As you might imagine, any solution geared towards improving the security of an application is going to be complex, consisting of many different moving parts. Thus, forcing customers to hunt for the components they need within your product data sheets and overviews is not going to be an effective way to convince those customers that you have a solution they might be in the market for. So what can vendors do to convince customers that they have a solution worth that customer’s time to evaluate? For starters, they can bundle various features into use cases that can be easily demonstrated to, evaluated, and consumed by customers. Along those lines, what would a bundle around the popular application security protection use case look like? While not an exhaustive list, here are some thoughts: - **App Proxy:** Putting a proxy in front of applications is perhaps one of the most basic application security requirements, and for good reason. Having an intermediary allows us to inspect and monitor traffic going to and from the application, as well as to block or filter as necessary for security purposes. - **Rate Limiting and Fast Access Control Lists (ACLs):** Flooding a site is an old standby of attackers. It is a primitive, yet effective tactic. Rate limiting is a relatively straightforward way to prevent this type of attack. Similarly, fast-performing Access Control Lists (ACLs) are another effective way to keep unwanted traffic at bay. - **Path Discovery:** Applying machine learning (ML) to traffic transiting the environment allows us to track the rate of requests, the identity of clients accessing applications, the size of the payloads being sent, and other important telemetry elements. Using ML allows us to identify and block nefarious traffic before it becomes a more serious issue – often in minutes as opposed to hours. - **Web Application Firewall:** WAF has become a required technology for application providers and should be included as a part of any application security bundle. - **L3/L4/L7 DDoS:** DDoS protection has also become a requirement for application providers and should also be included as part of any application security bundle. - **Bot Defense:** Advanced bots that know how to get around the defenses listed above can cause application providers monetary loss and reputation damage. As such, bot defense should also be included as part of an application security bundle. - **Auto-Certificates:** Speed of deploying applications is essential for remaining competitive, as is speed of protecting those applications. The ability to auto-issue certificates and to auto-register DNS for resources saves time, allowing application providers to go from no protection to full protection in a matter of minutes. - **Malicious User Detection:** Another great application for machine learning (ML) is quickly understanding which users and patterns appear to be behaving maliciously. This is something that often takes application providers hours or days to identify. With ML, this can be done in minutes, allowing those application providers to quickly take action and block/mitigate. - **Client-Side Defense:** Visibility into the end-user environment is something many application providers lack. The ability to inspect how JavaScript is being called, where requests are going, and what third party scripts are being called gives important insight that is extremely helpful for application security purposes. - **URI Routing:** The ability to quickly and easily control where certain requests are routing gives application providers the ability to block/control specific endpoints (URIs). No application security solution would be complete without this important feature. - **Service Policies:** Quick and easy policy deployment is a must for application security. The ability to chain together service policies as needed based on requirements, along with the ability to generate custom rules for steering traffic or allowing/denying traffic beyond the capabilities of the other defensive capabilities is another essential part of the total application security package. - **Synthetic Monitors:** How are applications performing externally? What are my customers experiencing? These are important questions that synthetic monitors allow a business to answer, which can quickly identify any issues that might affect the application. - **TLS Fingerprinting and Device Identification:** While IP addresses change frequently, TLS fingerprints and device identifiers change much more rarely. Thus, basing policies and rules on them rather than IP address makes a lot of sense when it comes to application security. - **Cross-Site Request Forgery Protection:** Scripts that operate cross-site can cause serious problems for application providers. Thus mitigating the risk they present should be part of any application security bundle as well. Securing applications is a top priority for nearly all businesses. While there are many routes to application security, bundles that allow security teams to quickly and easily secure applications and affect security posture in a self-service manner are becoming increasingly popular. These bundles inform application providers and allow them to make better, more informed decisions to improve security posture without introducing unnecessary friction to the end-user. The post [Application Security Protection for the Masses](https://www.securityweek.com/application-security-protection-for-the-masses/) appeared first on [SecurityWeek](https://www.securityweek.com). <https://www.securityweek.com/application-security-protection-for-the-masses/>
non_test
application security protection for the masses i’ve always found it entertaining that so many sales pitches are essentially a listing of features for the product or service being sold the reason i find this entertaining is that for anyone who has worked on the customer side or has ever listened to customers it is obvious that customers buy solutions not products thus the notion of showing off how proud you are of your product by rattling off a laundry list of features has always seemed a bit odd to me in other words customers have a number of different problems issues and challenges that they are looking to solve they are not necessarily interested in all of the different things your product or service can do rather they are interested in learning how your solution can help them address their strategic priorities and move forward on the goals they have set for their security and fraud problems it is incumbent upon vendors to understand that and to make it easy for potential customers to understand that mapping along those lines improving application security is a common goal customers have as you might imagine any solution geared towards improving the security of an application is going to be complex consisting of many different moving parts thus forcing customers to hunt for the components they need within your product data sheets and overviews is not going to be an effective way to convince those customers that you have a solution they might be in the market for so what can vendors do to convince customers that they have a solution worth that customer’s time to evaluate for starters they can bundle various features into use cases that can be easily demonstrated to evaluated and consumed by customers along those lines what would a bundle around the popular application security protection use case look like while not an exhaustive list here are some thoughts app proxy putting a proxy in front of applications is perhaps one of the most basic application security requirements and for good reason having an intermediary allows us to inspect and monitor traffic going to and from the application as well as to block or filter as necessary for security purposes rate limiting and fast access control lists acls flooding a site is an old standby of attackers it is a primitive yet effective tactic rate limiting is a relatively straightforward way to prevent this type of attack similarly fast performing access control lists acls are another effective way to keep unwanted traffic at bay path discovery applying machine learning ml to traffic transiting the environment allows us to track the rate of requests the identity of clients accessing applications the size of the payloads being sent and other important telemetry elements using ml allows us to identify and block nefarious traffic before it becomes a more serious issue – often in minutes as opposed to hours web application firewall waf has become a required technology for application providers and should be included as a part of any application security bundle ddos ddos protection has also become a requirement for application providers and should also be included as part of any application security bundle bot defense advanced bots that know how to get around the defenses listed above can cause application providers monetary loss and reputation damage as such bot defense should also be included as part of an application security bundle auto certificates speed of deploying applications is essential for remaining competitive as is speed of protecting those applications the ability to auto issue certificates and to auto register dns for resources saves time allowing application providers to go from no protection to full protection in a matter of minutes malicious user detection another great application for machine learning ml is quickly understanding which users and patterns appear to be behaving maliciously this is something that often takes application providers hours or days to identify with ml this can be done in minutes allowing those application providers to quickly take action and block mitigate client side defense visibility into the end user environment is something many application providers lack the ability to inspect how javascript is being called where requests are going and what third party scripts are being called gives important insight that is extremely helpful for application security purposes uri routing the ability to quickly and easily control where certain requests are routing gives application providers the ability to block control specific endpoints uris no application security solution would be complete without this important feature service policies quick and easy policy deployment is a must for application security the ability to chain together service policies as needed based on requirements along with the ability to generate custom rules for steering traffic or allowing denying traffic beyond the capabilities of the other defensive capabilities is another essential part of the total application security package synthetic monitors how are applications performing externally what are my customers experiencing these are important questions that synthetic monitors allow a business to answer which can quickly identify any issues that might affect the application tls fingerprinting and device identification while ip addresses change frequently tls fingerprints and device identifiers change much more rarely thus basing policies and rules on them rather than ip address makes a lot of sense when it comes to application security cross site request forgery protection scripts that operate cross site can cause serious problems for application providers thus mitigating the risk they present should be part of any application security bundle as well securing applications is a top priority for nearly all businesses while there are many routes to application security bundles that allow security teams to quickly and easily secure applications and affect security posture in a self service manner are becoming increasingly popular these bundles inform application providers and allow them to make better more informed decisions to improve security posture without introducing unnecessary friction to the end user the post appeared first on
0
116,790
9,884,089,418
IssuesEvent
2019-06-24 21:06:26
rancher/rio
https://api.github.com/repos/rancher/rio
closed
registry does not start due to cert-checker error
to-test
rio install against AKS 1.13.5 registry deployment stays at 0/2; error in the cert-checker init container is: Can't open /etc/registry/tls.crt for reading, No such file or directory 140601283124072:error:02001002:system library:fopen:No such file or directory:crypto/bio/bss_file.c:72:fopen('/etc/registry/tls.crt','r') 140601283124072:error:2006D080:BIO routines:BIO_new_file:no such file:crypto/bio/bss_file.c:79: unable to load certificate
1.0
registry does not start due to cert-checker error - rio install against AKS 1.13.5 registry deployment stays at 0/2; error in the cert-checker init container is: Can't open /etc/registry/tls.crt for reading, No such file or directory 140601283124072:error:02001002:system library:fopen:No such file or directory:crypto/bio/bss_file.c:72:fopen('/etc/registry/tls.crt','r') 140601283124072:error:2006D080:BIO routines:BIO_new_file:no such file:crypto/bio/bss_file.c:79: unable to load certificate
test
registry does not start due to cert checker error rio install against aks registry deployment stays at error in the cert checker init container is can t open etc registry tls crt for reading no such file or directory error system library fopen no such file or directory crypto bio bss file c fopen etc registry tls crt r error bio routines bio new file no such file crypto bio bss file c unable to load certificate
1
561,754
16,623,252,492
IssuesEvent
2021-06-03 06:09:36
rancher/rke2
https://api.github.com/repos/rancher/rke2
closed
Restoring from snapshot fails executable file not found in $PATH
kind/bug priority/critical-urgent
**Environmental Info:** RKE2 Version: * `v1.21.1-rc1+rke2r1` Node(s) CPU architecture, OS, and Version: Ubuntu 20.04 Cluster Configuration: Single node cluster **Describe the bug:** Restoring from snapshot fails with Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet **Steps To Reproduce:** Create a single node cluster Take snapshot Stop rke2 service Run killall Restore using the snapshot ``` rke2 server \ --cluster-reset \ --cluster-reset-restore-path=/var/lib/rancher/rke2/server/db/snapshots/on-demand-<unix-timestamp ``` **Expected behavior:** Reset should be run and cluster-reset msg is expected to be seen on the console **Actual behavior:** Errors with Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet and panic **Additional context / logs:** ``` sudo rke2 server --cluster-reset --cluster-reset-restore-path=/var/lib/rancher/rke2/server/db/snapshots/on-demand-ip-172-31-39-67-1621896967 WARN[0000] not running in CIS mode INFO[0000] starting kubelet to clean up old static pods INFO[0005] Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x4be89b] goroutine 1 [running]: os.(*Process).signal(0x0, 0x70cc4b8, 0x9be7958, 0x0, 0x0) /usr/local/go/src/os/exec_unix.go:63 +0x3b os.(*Process).Signal(...) /usr/local/go/src/os/exec.go:135 os.(*Process).kill(...) /usr/local/go/src/os/exec_posix.go:66 os.(*Process).Kill(...) /usr/local/go/src/os/exec.go:120 github.com/rancher/rke2/pkg/rke2.removeOldPodManifests(0x66d7963, 0x15, 0xc001699a70, 0x1, 0x0, 0x0) /source/pkg/rke2/rke2.go:229 +0x945 github.com/rancher/rke2/pkg/rke2.setup(0xc000a4b4a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /source/pkg/rke2/rke2.go:164 +0x8a5 github.com/rancher/rke2/pkg/rke2.Server(0xc000a4b4a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /source/pkg/rke2/rke2.go:48 +0x6b github.com/rancher/rke2/pkg/cli/cmds.ServerRun(0xc000a4b4a0, 0x0, 0x0) /source/pkg/cli/cmds/server.go:143 +0x98 github.com/urfave/cli.HandleAction(0x57f1f00, 0x68fece0, 0xc000a4b4a0, 0xc000a4b4a0, 0x0) /go/pkg/mod/github.com/urfave/cli@v1.22.5/app.go:524 +0x105 github.com/urfave/cli.Command.Run(0x6693772, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0x66da585, 0x15, 0xc0017b8a50, ...) /go/pkg/mod/github.com/urfave/cli@v1.22.5/command.go:173 +0x579 github.com/urfave/cli.(*App).Run(0xc0009afc00, 0xc000a52c00, 0x4, 0x4, 0x0, 0x0) /go/pkg/mod/github.com/urfave/cli@v1.22.5/app.go:277 +0x808 main.main() /source/main.go:20 +0x714 ```
1.0
Restoring from snapshot fails executable file not found in $PATH - **Environmental Info:** RKE2 Version: * `v1.21.1-rc1+rke2r1` Node(s) CPU architecture, OS, and Version: Ubuntu 20.04 Cluster Configuration: Single node cluster **Describe the bug:** Restoring from snapshot fails with Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet **Steps To Reproduce:** Create a single node cluster Take snapshot Stop rke2 service Run killall Restore using the snapshot ``` rke2 server \ --cluster-reset \ --cluster-reset-restore-path=/var/lib/rancher/rke2/server/db/snapshots/on-demand-<unix-timestamp ``` **Expected behavior:** Reset should be run and cluster-reset msg is expected to be seen on the console **Actual behavior:** Errors with Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet and panic **Additional context / logs:** ``` sudo rke2 server --cluster-reset --cluster-reset-restore-path=/var/lib/rancher/rke2/server/db/snapshots/on-demand-ip-172-31-39-67-1621896967 WARN[0000] not running in CIS mode INFO[0000] starting kubelet to clean up old static pods INFO[0005] Containerd Exited: exec: "containerd": executable file not found in $PATH, exiting kubelet panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x4be89b] goroutine 1 [running]: os.(*Process).signal(0x0, 0x70cc4b8, 0x9be7958, 0x0, 0x0) /usr/local/go/src/os/exec_unix.go:63 +0x3b os.(*Process).Signal(...) /usr/local/go/src/os/exec.go:135 os.(*Process).kill(...) /usr/local/go/src/os/exec_posix.go:66 os.(*Process).Kill(...) /usr/local/go/src/os/exec.go:120 github.com/rancher/rke2/pkg/rke2.removeOldPodManifests(0x66d7963, 0x15, 0xc001699a70, 0x1, 0x0, 0x0) /source/pkg/rke2/rke2.go:229 +0x945 github.com/rancher/rke2/pkg/rke2.setup(0xc000a4b4a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /source/pkg/rke2/rke2.go:164 +0x8a5 github.com/rancher/rke2/pkg/rke2.Server(0xc000a4b4a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /source/pkg/rke2/rke2.go:48 +0x6b github.com/rancher/rke2/pkg/cli/cmds.ServerRun(0xc000a4b4a0, 0x0, 0x0) /source/pkg/cli/cmds/server.go:143 +0x98 github.com/urfave/cli.HandleAction(0x57f1f00, 0x68fece0, 0xc000a4b4a0, 0xc000a4b4a0, 0x0) /go/pkg/mod/github.com/urfave/cli@v1.22.5/app.go:524 +0x105 github.com/urfave/cli.Command.Run(0x6693772, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0x66da585, 0x15, 0xc0017b8a50, ...) /go/pkg/mod/github.com/urfave/cli@v1.22.5/command.go:173 +0x579 github.com/urfave/cli.(*App).Run(0xc0009afc00, 0xc000a52c00, 0x4, 0x4, 0x0, 0x0) /go/pkg/mod/github.com/urfave/cli@v1.22.5/app.go:277 +0x808 main.main() /source/main.go:20 +0x714 ```
non_test
restoring from snapshot fails executable file not found in path environmental info version node s cpu architecture os and version ubuntu cluster configuration single node cluster describe the bug restoring from snapshot fails with containerd exited exec containerd executable file not found in path exiting kubelet steps to reproduce create a single node cluster take snapshot stop service run killall restore using the snapshot server cluster reset cluster reset restore path var lib rancher server db snapshots on demand unix timestamp expected behavior reset should be run and cluster reset msg is expected to be seen on the console actual behavior errors with containerd exited exec containerd executable file not found in path exiting kubelet and panic additional context logs sudo server cluster reset cluster reset restore path var lib rancher server db snapshots on demand ip warn not running in cis mode info starting kubelet to clean up old static pods info containerd exited exec containerd executable file not found in path exiting kubelet panic runtime error invalid memory address or nil pointer dereference goroutine os process signal usr local go src os exec unix go os process signal usr local go src os exec go os process kill usr local go src os exec posix go os process kill usr local go src os exec go github com rancher pkg removeoldpodmanifests source pkg go github com rancher pkg setup source pkg go github com rancher pkg server source pkg go github com rancher pkg cli cmds serverrun source pkg cli cmds server go github com urfave cli handleaction go pkg mod github com urfave cli app go github com urfave cli command run go pkg mod github com urfave cli command go github com urfave cli app run go pkg mod github com urfave cli app go main main source main go
0
72,488
3,386,466,912
IssuesEvent
2015-11-27 17:48:42
ccrama/Slide
https://api.github.com/repos/ccrama/Slide
opened
Slide reopens itself from background when leaved in landscape
bug high priority
Happens every minute or so, when it isn't killed through recents. Newest self compiled apk. @ccrama Please fix this before pushing newest update
1.0
Slide reopens itself from background when leaved in landscape - Happens every minute or so, when it isn't killed through recents. Newest self compiled apk. @ccrama Please fix this before pushing newest update
non_test
slide reopens itself from background when leaved in landscape happens every minute or so when it isn t killed through recents newest self compiled apk ccrama please fix this before pushing newest update
0
252,765
8,041,562,805
IssuesEvent
2018-07-31 03:45:11
Zicerite/Gavania-Project
https://api.github.com/repos/Zicerite/Gavania-Project
closed
Spiking skill for Swordsman
Medium Priority Skill Related
Swordsman should have a skill that spikes enemies downward. Deals decent damage, but not much else.
1.0
Spiking skill for Swordsman - Swordsman should have a skill that spikes enemies downward. Deals decent damage, but not much else.
non_test
spiking skill for swordsman swordsman should have a skill that spikes enemies downward deals decent damage but not much else
0
352,332
32,060,099,919
IssuesEvent
2023-09-24 15:01:12
tijlleenders/ZinZen
https://api.github.com/repos/tijlleenders/ZinZen
opened
Add Playwright Tests for sharing feature
UI devex test dev research more_than_a_day
Follow the steps: - - On User 1 window 1. Open app, finish onboarding 2. Select a goal on my goals page 3. click on circle and select share action 4. select "share privately" from the popup 5. click on plus btn to add contact 6. write a name "User 2" and click on share invitation - Open that link in new window that will be acting as User 2 1. Open app, finish onboarding 2. paste link and add the contact - On user 1 window 8. click on the contact btn again to share the goal this time - On user 2 window 4. refresh the window 5. switch to my time and switch back to my goals 6. check if inbox icon is visible 7. click on inbox and check whether goal is present or not
1.0
Add Playwright Tests for sharing feature - Follow the steps: - - On User 1 window 1. Open app, finish onboarding 2. Select a goal on my goals page 3. click on circle and select share action 4. select "share privately" from the popup 5. click on plus btn to add contact 6. write a name "User 2" and click on share invitation - Open that link in new window that will be acting as User 2 1. Open app, finish onboarding 2. paste link and add the contact - On user 1 window 8. click on the contact btn again to share the goal this time - On user 2 window 4. refresh the window 5. switch to my time and switch back to my goals 6. check if inbox icon is visible 7. click on inbox and check whether goal is present or not
test
add playwright tests for sharing feature follow the steps on user window open app finish onboarding select a goal on my goals page click on circle and select share action select share privately from the popup click on plus btn to add contact write a name user and click on share invitation open that link in new window that will be acting as user open app finish onboarding paste link and add the contact on user window click on the contact btn again to share the goal this time on user window refresh the window switch to my time and switch back to my goals check if inbox icon is visible click on inbox and check whether goal is present or not
1
323,159
27,700,042,837
IssuesEvent
2023-03-14 07:10:56
astronomer/astronomer-providers
https://api.github.com/repos/astronomer/astronomer-providers
closed
Test the timeout parameter for all the Triggers
testing
Test the timeout parameter for all the Triggers (see Trigger table in airflow )
1.0
Test the timeout parameter for all the Triggers - Test the timeout parameter for all the Triggers (see Trigger table in airflow )
test
test the timeout parameter for all the triggers test the timeout parameter for all the triggers see trigger table in airflow
1
96,445
8,614,954,380
IssuesEvent
2018-11-19 19:03:20
nebula-plugins/gradle-nebula-integration
https://api.github.com/repos/nebula-plugins/gradle-nebula-integration
closed
Dependencies which have transitive=false in ivy file are not included in alignment
priority testable
I have a dependency which has transitive dependencies defined in ivy file. All of them are members of the same group for alignment. However, one of them is defined with transitive=false. `<dependency org="com.netflix.governator" name="governator-jetty" rev="1.14.4" conf="compile-&gt;default" transitive="false"/>` This dependency is not added to a platform for alignment. `./gradlew dI --dependency governator --configuration compile` gives the following output. ``` > Task :dependencyInsight com.netflix.governator:governator-api:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-api:1.17.5 \--- com.netflix.governator:governator-core:1.17.5 +--- netflix:library:1.173 | \--- compile +--- com.netflix.governator:governator-jersey:1.17.5 | +--- netflix:library:1.173 (*) | \--- compile +--- com.netflix.governator:governator-servlet:1.17.5 | +--- netflix:library:1.173 (*) | \--- compile \--- compile com.netflix.governator:governator-core:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - Forced - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-core:1.17.5 +--- com.netflix.governator:governator-jersey:1.17.5 | +--- netflix:library:1.173 | | \--- compile | \--- compile \--- com.netflix.governator:governator-servlet:1.17.5 +--- netflix:library:1.173 (*) \--- compile com.netflix.governator:governator-core:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-jersey:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 - Forced com.netflix.governator:governator-jersey:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-jetty:1.14.4 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] com.netflix.governator:governator-jetty:1.14.4 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-providers:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-providers:1.17.5 \--- com.netflix.governator:governator-jersey:1.17.5 +--- netflix:library:1.173 | \--- compile \--- compile com.netflix.governator:governator-servlet:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 - Forced com.netflix.governator:governator-servlet:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile ``` See `com.netflix.governator:governator-jetty:1.14.4` Here is the full example project https://github.com/nebula-plugins/gradle-nebula-integration/tree/master/transitive-false-not-aligned
1.0
Dependencies which have transitive=false in ivy file are not included in alignment - I have a dependency which has transitive dependencies defined in ivy file. All of them are members of the same group for alignment. However, one of them is defined with transitive=false. `<dependency org="com.netflix.governator" name="governator-jetty" rev="1.14.4" conf="compile-&gt;default" transitive="false"/>` This dependency is not added to a platform for alignment. `./gradlew dI --dependency governator --configuration compile` gives the following output. ``` > Task :dependencyInsight com.netflix.governator:governator-api:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-api:1.17.5 \--- com.netflix.governator:governator-core:1.17.5 +--- netflix:library:1.173 | \--- compile +--- com.netflix.governator:governator-jersey:1.17.5 | +--- netflix:library:1.173 (*) | \--- compile +--- com.netflix.governator:governator-servlet:1.17.5 | +--- netflix:library:1.173 (*) | \--- compile \--- compile com.netflix.governator:governator-core:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - Forced - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-core:1.17.5 +--- com.netflix.governator:governator-jersey:1.17.5 | +--- netflix:library:1.173 | | \--- compile | \--- compile \--- com.netflix.governator:governator-servlet:1.17.5 +--- netflix:library:1.173 (*) \--- compile com.netflix.governator:governator-core:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-jersey:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 - Forced com.netflix.governator:governator-jersey:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-jetty:1.14.4 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] com.netflix.governator:governator-jetty:1.14.4 \--- netflix:library:1.173 \--- compile com.netflix.governator:governator-providers:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 com.netflix.governator:governator-providers:1.17.5 \--- com.netflix.governator:governator-jersey:1.17.5 +--- netflix:library:1.173 | \--- compile \--- compile com.netflix.governator:governator-servlet:1.17.5 variant "runtime" [ org.gradle.status = release (not requested) org.gradle.usage = java-runtime (not requested) org.gradle.component.category = library (not requested) ] Selection reasons: - By constraint : belongs to platform aligned-platform:governator:1.17.5 - Forced com.netflix.governator:governator-servlet:1.14.4 -> 1.17.5 \--- netflix:library:1.173 \--- compile ``` See `com.netflix.governator:governator-jetty:1.14.4` Here is the full example project https://github.com/nebula-plugins/gradle-nebula-integration/tree/master/transitive-false-not-aligned
test
dependencies which have transitive false in ivy file are not included in alignment i have a dependency which has transitive dependencies defined in ivy file all of them are members of the same group for alignment however one of them is defined with transitive false this dependency is not added to a platform for alignment gradlew di dependency governator configuration compile gives the following output task dependencyinsight com netflix governator governator api variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested selection reasons by constraint belongs to platform aligned platform governator com netflix governator governator api com netflix governator governator core netflix library compile com netflix governator governator jersey netflix library compile com netflix governator governator servlet netflix library compile compile com netflix governator governator core variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested selection reasons forced by constraint belongs to platform aligned platform governator com netflix governator governator core com netflix governator governator jersey netflix library compile compile com netflix governator governator servlet netflix library compile com netflix governator governator core netflix library compile com netflix governator governator jersey variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested selection reasons by constraint belongs to platform aligned platform governator forced com netflix governator governator jersey netflix library compile com netflix governator governator jetty variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested com netflix governator governator jetty netflix library compile com netflix governator governator providers variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested selection reasons by constraint belongs to platform aligned platform governator com netflix governator governator providers com netflix governator governator jersey netflix library compile compile com netflix governator governator servlet variant runtime org gradle status release not requested org gradle usage java runtime not requested org gradle component category library not requested selection reasons by constraint belongs to platform aligned platform governator forced com netflix governator governator servlet netflix library compile see com netflix governator governator jetty here is the full example project
1
297,585
9,178,830,844
IssuesEvent
2019-03-05 00:35:27
minio/mc
https://api.github.com/repos/minio/mc
closed
Change --insecure flag to --skipcertcheck
priority: medium
*** Not using standard template *** `mc` currently uses `--insecureflag` flag for the case where self signed certificates are used. The name is a little misleading. So, this needs to be changed to `--skipcertcheck`
1.0
Change --insecure flag to --skipcertcheck - *** Not using standard template *** `mc` currently uses `--insecureflag` flag for the case where self signed certificates are used. The name is a little misleading. So, this needs to be changed to `--skipcertcheck`
non_test
change insecure flag to skipcertcheck not using standard template mc currently uses insecureflag flag for the case where self signed certificates are used the name is a little misleading so this needs to be changed to skipcertcheck
0
245,623
20,789,934,698
IssuesEvent
2022-03-17 00:08:35
Ale-Torres/BrowserQuest
https://api.github.com/repos/Ale-Torres/BrowserQuest
closed
Need to add unit tests for astar.js
no unit testing
We need to add unit tests to astar.js so that when we fix the code smell, we can ensure it is still working correctly
1.0
Need to add unit tests for astar.js - We need to add unit tests to astar.js so that when we fix the code smell, we can ensure it is still working correctly
test
need to add unit tests for astar js we need to add unit tests to astar js so that when we fix the code smell we can ensure it is still working correctly
1
179,867
13,908,229,502
IssuesEvent
2020-10-20 13:34:42
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
app-linux-pdf-converter v2.1.10 (r4.1)
r4.1-bullseye-cur-test r4.1-buster-cur-test r4.1-centos8-cur-test r4.1-dom0-cur-test r4.1-fc29-cur-test r4.1-fc30-cur-test r4.1-fc31-cur-test r4.1-fc32-cur-test
Update of app-linux-pdf-converter to v2.1.10 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-pdf-converter/commit/0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-pdf-converter/compare/v2.1.8...v2.1.10): QubesOS/qubes-app-linux-pdf-converter@0c9d97e version 2.1.10 QubesOS/qubes-app-linux-pdf-converter@6b7f71f Python38 is now available for CentOS8 QubesOS/qubes-app-linux-pdf-converter@6d1ba08 version 2.1.9 QubesOS/qubes-app-linux-pdf-converter@b83e354 require python 3.7+, exclude stretch, jessie and centos QubesOS/qubes-app-linux-pdf-converter@653c55f Use proper python package installation for the VM part too QubesOS/qubes-app-linux-pdf-converter@229e68a Drop executable bit from nautilus extension QubesOS/qubes-app-linux-pdf-converter@617c2a7 Add qubespdfconverter/__init__.py QubesOS/qubes-app-linux-pdf-converter@6d457d6 Move tests back to qubespdfconverter.tests QubesOS/qubes-app-linux-pdf-converter@f1559c9 Install only tests with older python QubesOS/qubes-app-linux-pdf-converter@04991c3 travis: include R4.0 dom0 too QubesOS/qubes-app-linux-pdf-converter@73a9012 Adjust dependencies after python rewrite QubesOS/qubes-app-linux-pdf-converter@60b6b5c Merge remote-tracking branch 'origin/pr/7' QubesOS/qubes-app-linux-pdf-converter@1ee08f7 makefile: Resolve makefile conflict QubesOS/qubes-app-linux-pdf-converter@64cc14f pylint: Add expression-not-assigned QubesOS/qubes-app-linux-pdf-converter@379659b client: Add support for older tqdm versions QubesOS/qubes-app-linux-pdf-converter@f1d35f2 client: Handle out of space error QubesOS/qubes-app-linux-pdf-converter@5eab363 client: Fix output spacing QubesOS/qubes-app-linux-pdf-converter@c87e61b client: Simplify image appending QubesOS/qubes-app-linux-pdf-converter@9b655af pylint: Add bad-continuation to .pylintrc QubesOS/qubes-app-linux-pdf-converter@388278a debian: update control syntax for python nautilus QubesOS/qubes-app-linux-pdf-converter@2abacda meta: Copyright info QubesOS/qubes-app-linux-pdf-converter@ef41913 client: Exit with 1 on error QubesOS/qubes-app-linux-pdf-converter@bcbaf9e client: Implement bulk saves and remove reps appropriately QubesOS/qubes-app-linux-pdf-converter@8e32abb server: Rename batch entry variables QubesOS/qubes-app-linux-pdf-converter@c00e7a1 server: Prevent CPU hogging by PNG tasks QubesOS/qubes-app-linux-pdf-converter@5772fdf server: Handle non-ASCII output from pdfinfo(1) QubesOS/qubes-app-linux-pdf-converter@61437ca client: Use python3 shebang QubesOS/qubes-app-linux-pdf-converter@deb6c87 travis: Update imports and pylint exceptions QubesOS/qubes-app-linux-pdf-converter@467a7c3 travis: remove --exit-code QubesOS/qubes-app-linux-pdf-converter@0b1e80c travis: Attempt to fix pylint QubesOS/qubes-app-linux-pdf-converter@cd08b7e debian: fix typo in rules QubesOS/qubes-app-linux-pdf-converter@4e3fee3 travis: Re-add --exit-code because i was right the first time QubesOS/qubes-app-linux-pdf-converter@2718653 travis: fix paths and try without --exit-code QubesOS/qubes-app-linux-pdf-converter@32dd0d2 travis: Add pylint dependency QubesOS/qubes-app-linux-pdf-converter@35524b7 client, server: Replace shellcheck with pylint QubesOS/qubes-app-linux-pdf-converter@4cfb5e9 client: Update docstring QubesOS/qubes-app-linux-pdf-converter@5ac1bf0 setup: Add tqdm dependency QubesOS/qubes-app-linux-pdf-converter@21a6314 client: Update shebang to 3.7 QubesOS/qubes-app-linux-pdf-converter@f05f4d5 client: Fix lower bound for batch size QubesOS/qubes-app-linux-pdf-converter@7243999 client, server: Update UI, design, and error handling QubesOS/qubes-app-linux-pdf-converter@924824f client: Replace namedtuple with dataclass QubesOS/qubes-app-linux-pdf-converter@2a5d8b9 client: Add initial version of tqdm output QubesOS/qubes-app-linux-pdf-converter@9a61502 client: Implement --in-place QubesOS/qubes-app-linux-pdf-converter@225780e client: Add --archive, --dry-run, and --in-place QubesOS/qubes-app-linux-pdf-converter@24dc64f client, server: Rename unsanitized file class QubesOS/qubes-app-linux-pdf-converter@4e6b286 client, server: Remove unused imports QubesOS/qubes-app-linux-pdf-converter@3fd4636 server: Remove nonexistent PageError QubesOS/qubes-app-linux-pdf-converter@4e5e87e client, server: Add Representation & *File objects QubesOS/qubes-app-linux-pdf-converter@dc4742b client: Replace click.echo() QubesOS/qubes-app-linux-pdf-converter@7c9a21f server: Update sending process QubesOS/qubes-app-linux-pdf-converter@1fa078b server: Update error handling QubesOS/qubes-app-linux-pdf-converter@06bafaf server: Remove recv_pdf() QubesOS/qubes-app-linux-pdf-converter@89ebda3 client: Update error handling QubesOS/qubes-app-linux-pdf-converter@8c0054c client: Add CLI parameter handling QubesOS/qubes-app-linux-pdf-converter@8905550 client: Start batch processing code QubesOS/qubes-app-linux-pdf-converter@11d610a Rename source directory QubesOS/qubes-app-linux-pdf-converter@8dd8d6f server: Remove uneeded quoting in identify command QubesOS/qubes-app-linux-pdf-converter@523aab8 client: Remove unused recv() QubesOS/qubes-app-linux-pdf-converter@1c4f61d client: Remove unneeded PID prefix for tmpdir QubesOS/qubes-app-linux-pdf-converter@e057643 client: Remove extra byte for send() QubesOS/qubes-app-linux-pdf-converter@ee7420d client: Fix error handling for readexactly() QubesOS/qubes-app-linux-pdf-converter@d7dade5 meta: Remove wrapper QubesOS/qubes-app-linux-pdf-converter@a0cffae meta: Remove extra file QubesOS/qubes-app-linux-pdf-converter@456ac9a tests: Moved to pdf-converter/tests QubesOS/qubes-app-linux-pdf-converter@23a45e5 meta: Remove extra file QubesOS/qubes-app-linux-pdf-converter@5979f9f client, server: Move to pdf-converter directory QubesOS/qubes-app-linux-pdf-converter@f4291c3 client, server: Update to asyncio and new model QubesOS/qubes-app-linux-pdf-converter@4d1ac9e client, server: Update initial phase QubesOS/qubes-app-linux-pdf-converter@f66da6e client: Combine image dimension functions QubesOS/qubes-app-linux-pdf-converter@8178f1d client, server: Update utility functions QubesOS/qubes-app-linux-pdf-converter@c6f6be4 client, server: Add custom exceptions QubesOS/qubes-app-linux-pdf-converter@7913920 client: Remove redundant check condition QubesOS/qubes-app-linux-pdf-converter@672ad1a client, server: Replace recv() with recvline_b() QubesOS/qubes-app-linux-pdf-converter@3fd5912 Update docs QubesOS/qubes-app-linux-pdf-converter@54e25fe Add support for multiple files QubesOS/qubes-app-linux-pdf-converter@574e63c readme: Update language QubesOS/qubes-app-linux-pdf-converter@a4b70cb server, client: Remove unecessary documentation QubesOS/qubes-app-linux-pdf-converter@7279415 wrapper, client, server: Qualify variable names QubesOS/qubes-app-linux-pdf-converter@88acf01 wrapper: Add non-file check QubesOS/qubes-app-linux-pdf-converter@098ef25 wrapper: Simplify argument parsing QubesOS/qubes-app-linux-pdf-converter@e43c071 server: Fix location of STDOUT flush call QubesOS/qubes-app-linux-pdf-converter@82b7e73 client: Use OG stdout file descriptor QubesOS/qubes-app-linux-pdf-converter@dd4abde server, client: Grammar QubesOS/qubes-app-linux-pdf-converter@f937480 server: Avoid errors if subprocess command fails QubesOS/qubes-app-linux-pdf-converter@8d5bc16 client: Handle EOFErrors if server suddenly dies QubesOS/qubes-app-linux-pdf-converter@2857aa4 client, server: Fix stream flushing QubesOS/qubes-app-linux-pdf-converter@c6831f5 wrapper: Move logging config to global namespace QubesOS/qubes-app-linux-pdf-converter@3d941e8 wrapper: Trim down argument parsing QubesOS/qubes-app-linux-pdf-converter@21407f1 wrapper: Fix leftover debugging command QubesOS/qubes-app-linux-pdf-converter@150793c readme: Update usage example QubesOS/qubes-app-linux-pdf-converter@b8b41da Add .py extensions and update install commands QubesOS/qubes-app-linux-pdf-converter@ec754fc wrapper: Add documentation QubesOS/qubes-app-linux-pdf-converter@5cfaab7 server: Make check_output() return a String instead of a bytestring QubesOS/qubes-app-linux-pdf-converter@8a8283d client, server: Add documentation QubesOS/qubes-app-linux-pdf-converter@f6c90fb server: Refactor to Python 3 QubesOS/qubes-app-linux-pdf-converter@552d2f4 client: PEP 8 QubesOS/qubes-app-linux-pdf-converter@516cd21 client: Add qrexec wrapper functions QubesOS/qubes-app-linux-pdf-converter@588a9e2 wrapper: PEP 8 QubesOS/qubes-app-linux-pdf-converter@4d88ecf wrapper, client: Re-add logging QubesOS/qubes-app-linux-pdf-converter@0a7cfae client: Update to Python 3 QubesOS/qubes-app-linux-pdf-converter@fd9175f wrapper: Remove logging QubesOS/qubes-app-linux-pdf-converter@a014fb3 wrapper: Remove unneeded main() try block QubesOS/qubes-app-linux-pdf-converter@9668bfb wrapper: Prepare for multiple file support QubesOS/qubes-app-linux-pdf-converter@7ef5b33 wrapper: Add logging and trim options QubesOS/qubes-app-linux-pdf-converter@4e0d635 wrapper: Update qvm-convert-pdf into Python 3 QubesOS/qubes-app-linux-pdf-converter@a84a215 readme: Remove extra parenthesis QubesOS/qubes-app-linux-pdf-converter@99629af Merge remote-tracking branch 'origin/pr/10' QubesOS/qubes-app-linux-pdf-converter@f748f50 debian: conditional python version dependencies Referenced issues: QubesOS/qubes-issues#5868 If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 current repo` (available 7 days from now) * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
8.0
app-linux-pdf-converter v2.1.10 (r4.1) - Update of app-linux-pdf-converter to v2.1.10 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-pdf-converter/commit/0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-pdf-converter/compare/v2.1.8...v2.1.10): QubesOS/qubes-app-linux-pdf-converter@0c9d97e version 2.1.10 QubesOS/qubes-app-linux-pdf-converter@6b7f71f Python38 is now available for CentOS8 QubesOS/qubes-app-linux-pdf-converter@6d1ba08 version 2.1.9 QubesOS/qubes-app-linux-pdf-converter@b83e354 require python 3.7+, exclude stretch, jessie and centos QubesOS/qubes-app-linux-pdf-converter@653c55f Use proper python package installation for the VM part too QubesOS/qubes-app-linux-pdf-converter@229e68a Drop executable bit from nautilus extension QubesOS/qubes-app-linux-pdf-converter@617c2a7 Add qubespdfconverter/__init__.py QubesOS/qubes-app-linux-pdf-converter@6d457d6 Move tests back to qubespdfconverter.tests QubesOS/qubes-app-linux-pdf-converter@f1559c9 Install only tests with older python QubesOS/qubes-app-linux-pdf-converter@04991c3 travis: include R4.0 dom0 too QubesOS/qubes-app-linux-pdf-converter@73a9012 Adjust dependencies after python rewrite QubesOS/qubes-app-linux-pdf-converter@60b6b5c Merge remote-tracking branch 'origin/pr/7' QubesOS/qubes-app-linux-pdf-converter@1ee08f7 makefile: Resolve makefile conflict QubesOS/qubes-app-linux-pdf-converter@64cc14f pylint: Add expression-not-assigned QubesOS/qubes-app-linux-pdf-converter@379659b client: Add support for older tqdm versions QubesOS/qubes-app-linux-pdf-converter@f1d35f2 client: Handle out of space error QubesOS/qubes-app-linux-pdf-converter@5eab363 client: Fix output spacing QubesOS/qubes-app-linux-pdf-converter@c87e61b client: Simplify image appending QubesOS/qubes-app-linux-pdf-converter@9b655af pylint: Add bad-continuation to .pylintrc QubesOS/qubes-app-linux-pdf-converter@388278a debian: update control syntax for python nautilus QubesOS/qubes-app-linux-pdf-converter@2abacda meta: Copyright info QubesOS/qubes-app-linux-pdf-converter@ef41913 client: Exit with 1 on error QubesOS/qubes-app-linux-pdf-converter@bcbaf9e client: Implement bulk saves and remove reps appropriately QubesOS/qubes-app-linux-pdf-converter@8e32abb server: Rename batch entry variables QubesOS/qubes-app-linux-pdf-converter@c00e7a1 server: Prevent CPU hogging by PNG tasks QubesOS/qubes-app-linux-pdf-converter@5772fdf server: Handle non-ASCII output from pdfinfo(1) QubesOS/qubes-app-linux-pdf-converter@61437ca client: Use python3 shebang QubesOS/qubes-app-linux-pdf-converter@deb6c87 travis: Update imports and pylint exceptions QubesOS/qubes-app-linux-pdf-converter@467a7c3 travis: remove --exit-code QubesOS/qubes-app-linux-pdf-converter@0b1e80c travis: Attempt to fix pylint QubesOS/qubes-app-linux-pdf-converter@cd08b7e debian: fix typo in rules QubesOS/qubes-app-linux-pdf-converter@4e3fee3 travis: Re-add --exit-code because i was right the first time QubesOS/qubes-app-linux-pdf-converter@2718653 travis: fix paths and try without --exit-code QubesOS/qubes-app-linux-pdf-converter@32dd0d2 travis: Add pylint dependency QubesOS/qubes-app-linux-pdf-converter@35524b7 client, server: Replace shellcheck with pylint QubesOS/qubes-app-linux-pdf-converter@4cfb5e9 client: Update docstring QubesOS/qubes-app-linux-pdf-converter@5ac1bf0 setup: Add tqdm dependency QubesOS/qubes-app-linux-pdf-converter@21a6314 client: Update shebang to 3.7 QubesOS/qubes-app-linux-pdf-converter@f05f4d5 client: Fix lower bound for batch size QubesOS/qubes-app-linux-pdf-converter@7243999 client, server: Update UI, design, and error handling QubesOS/qubes-app-linux-pdf-converter@924824f client: Replace namedtuple with dataclass QubesOS/qubes-app-linux-pdf-converter@2a5d8b9 client: Add initial version of tqdm output QubesOS/qubes-app-linux-pdf-converter@9a61502 client: Implement --in-place QubesOS/qubes-app-linux-pdf-converter@225780e client: Add --archive, --dry-run, and --in-place QubesOS/qubes-app-linux-pdf-converter@24dc64f client, server: Rename unsanitized file class QubesOS/qubes-app-linux-pdf-converter@4e6b286 client, server: Remove unused imports QubesOS/qubes-app-linux-pdf-converter@3fd4636 server: Remove nonexistent PageError QubesOS/qubes-app-linux-pdf-converter@4e5e87e client, server: Add Representation & *File objects QubesOS/qubes-app-linux-pdf-converter@dc4742b client: Replace click.echo() QubesOS/qubes-app-linux-pdf-converter@7c9a21f server: Update sending process QubesOS/qubes-app-linux-pdf-converter@1fa078b server: Update error handling QubesOS/qubes-app-linux-pdf-converter@06bafaf server: Remove recv_pdf() QubesOS/qubes-app-linux-pdf-converter@89ebda3 client: Update error handling QubesOS/qubes-app-linux-pdf-converter@8c0054c client: Add CLI parameter handling QubesOS/qubes-app-linux-pdf-converter@8905550 client: Start batch processing code QubesOS/qubes-app-linux-pdf-converter@11d610a Rename source directory QubesOS/qubes-app-linux-pdf-converter@8dd8d6f server: Remove uneeded quoting in identify command QubesOS/qubes-app-linux-pdf-converter@523aab8 client: Remove unused recv() QubesOS/qubes-app-linux-pdf-converter@1c4f61d client: Remove unneeded PID prefix for tmpdir QubesOS/qubes-app-linux-pdf-converter@e057643 client: Remove extra byte for send() QubesOS/qubes-app-linux-pdf-converter@ee7420d client: Fix error handling for readexactly() QubesOS/qubes-app-linux-pdf-converter@d7dade5 meta: Remove wrapper QubesOS/qubes-app-linux-pdf-converter@a0cffae meta: Remove extra file QubesOS/qubes-app-linux-pdf-converter@456ac9a tests: Moved to pdf-converter/tests QubesOS/qubes-app-linux-pdf-converter@23a45e5 meta: Remove extra file QubesOS/qubes-app-linux-pdf-converter@5979f9f client, server: Move to pdf-converter directory QubesOS/qubes-app-linux-pdf-converter@f4291c3 client, server: Update to asyncio and new model QubesOS/qubes-app-linux-pdf-converter@4d1ac9e client, server: Update initial phase QubesOS/qubes-app-linux-pdf-converter@f66da6e client: Combine image dimension functions QubesOS/qubes-app-linux-pdf-converter@8178f1d client, server: Update utility functions QubesOS/qubes-app-linux-pdf-converter@c6f6be4 client, server: Add custom exceptions QubesOS/qubes-app-linux-pdf-converter@7913920 client: Remove redundant check condition QubesOS/qubes-app-linux-pdf-converter@672ad1a client, server: Replace recv() with recvline_b() QubesOS/qubes-app-linux-pdf-converter@3fd5912 Update docs QubesOS/qubes-app-linux-pdf-converter@54e25fe Add support for multiple files QubesOS/qubes-app-linux-pdf-converter@574e63c readme: Update language QubesOS/qubes-app-linux-pdf-converter@a4b70cb server, client: Remove unecessary documentation QubesOS/qubes-app-linux-pdf-converter@7279415 wrapper, client, server: Qualify variable names QubesOS/qubes-app-linux-pdf-converter@88acf01 wrapper: Add non-file check QubesOS/qubes-app-linux-pdf-converter@098ef25 wrapper: Simplify argument parsing QubesOS/qubes-app-linux-pdf-converter@e43c071 server: Fix location of STDOUT flush call QubesOS/qubes-app-linux-pdf-converter@82b7e73 client: Use OG stdout file descriptor QubesOS/qubes-app-linux-pdf-converter@dd4abde server, client: Grammar QubesOS/qubes-app-linux-pdf-converter@f937480 server: Avoid errors if subprocess command fails QubesOS/qubes-app-linux-pdf-converter@8d5bc16 client: Handle EOFErrors if server suddenly dies QubesOS/qubes-app-linux-pdf-converter@2857aa4 client, server: Fix stream flushing QubesOS/qubes-app-linux-pdf-converter@c6831f5 wrapper: Move logging config to global namespace QubesOS/qubes-app-linux-pdf-converter@3d941e8 wrapper: Trim down argument parsing QubesOS/qubes-app-linux-pdf-converter@21407f1 wrapper: Fix leftover debugging command QubesOS/qubes-app-linux-pdf-converter@150793c readme: Update usage example QubesOS/qubes-app-linux-pdf-converter@b8b41da Add .py extensions and update install commands QubesOS/qubes-app-linux-pdf-converter@ec754fc wrapper: Add documentation QubesOS/qubes-app-linux-pdf-converter@5cfaab7 server: Make check_output() return a String instead of a bytestring QubesOS/qubes-app-linux-pdf-converter@8a8283d client, server: Add documentation QubesOS/qubes-app-linux-pdf-converter@f6c90fb server: Refactor to Python 3 QubesOS/qubes-app-linux-pdf-converter@552d2f4 client: PEP 8 QubesOS/qubes-app-linux-pdf-converter@516cd21 client: Add qrexec wrapper functions QubesOS/qubes-app-linux-pdf-converter@588a9e2 wrapper: PEP 8 QubesOS/qubes-app-linux-pdf-converter@4d88ecf wrapper, client: Re-add logging QubesOS/qubes-app-linux-pdf-converter@0a7cfae client: Update to Python 3 QubesOS/qubes-app-linux-pdf-converter@fd9175f wrapper: Remove logging QubesOS/qubes-app-linux-pdf-converter@a014fb3 wrapper: Remove unneeded main() try block QubesOS/qubes-app-linux-pdf-converter@9668bfb wrapper: Prepare for multiple file support QubesOS/qubes-app-linux-pdf-converter@7ef5b33 wrapper: Add logging and trim options QubesOS/qubes-app-linux-pdf-converter@4e0d635 wrapper: Update qvm-convert-pdf into Python 3 QubesOS/qubes-app-linux-pdf-converter@a84a215 readme: Remove extra parenthesis QubesOS/qubes-app-linux-pdf-converter@99629af Merge remote-tracking branch 'origin/pr/10' QubesOS/qubes-app-linux-pdf-converter@f748f50 debian: conditional python version dependencies Referenced issues: QubesOS/qubes-issues#5868 If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 current repo` (available 7 days from now) * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-pdf-converter 0c9d97e0b8cef2d4ce2beed65e2c64d70e235de8 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
test
app linux pdf converter update of app linux pdf converter to for qubes see comments below for details built from qubesos qubes app linux pdf converter version qubesos qubes app linux pdf converter is now available for qubesos qubes app linux pdf converter version qubesos qubes app linux pdf converter require python exclude stretch jessie and centos qubesos qubes app linux pdf converter use proper python package installation for the vm part too qubesos qubes app linux pdf converter drop executable bit from nautilus extension qubesos qubes app linux pdf converter add qubespdfconverter init py qubesos qubes app linux pdf converter move tests back to qubespdfconverter tests qubesos qubes app linux pdf converter install only tests with older python qubesos qubes app linux pdf converter travis include too qubesos qubes app linux pdf converter adjust dependencies after python rewrite qubesos qubes app linux pdf converter merge remote tracking branch origin pr qubesos qubes app linux pdf converter makefile resolve makefile conflict qubesos qubes app linux pdf converter pylint add expression not assigned qubesos qubes app linux pdf converter client add support for older tqdm versions qubesos qubes app linux pdf converter client handle out of space error qubesos qubes app linux pdf converter client fix output spacing qubesos qubes app linux pdf converter client simplify image appending qubesos qubes app linux pdf converter pylint add bad continuation to pylintrc qubesos qubes app linux pdf converter debian update control syntax for python nautilus qubesos qubes app linux pdf converter meta copyright info qubesos qubes app linux pdf converter client exit with on error qubesos qubes app linux pdf converter client implement bulk saves and remove reps appropriately qubesos qubes app linux pdf converter server rename batch entry variables qubesos qubes app linux pdf converter server prevent cpu hogging by png tasks qubesos qubes app linux pdf converter server handle non ascii output from pdfinfo qubesos qubes app linux pdf converter client use shebang qubesos qubes app linux pdf converter travis update imports and pylint exceptions qubesos qubes app linux pdf converter travis remove exit code qubesos qubes app linux pdf converter travis attempt to fix pylint qubesos qubes app linux pdf converter debian fix typo in rules qubesos qubes app linux pdf converter travis re add exit code because i was right the first time qubesos qubes app linux pdf converter travis fix paths and try without exit code qubesos qubes app linux pdf converter travis add pylint dependency qubesos qubes app linux pdf converter client server replace shellcheck with pylint qubesos qubes app linux pdf converter client update docstring qubesos qubes app linux pdf converter setup add tqdm dependency qubesos qubes app linux pdf converter client update shebang to qubesos qubes app linux pdf converter client fix lower bound for batch size qubesos qubes app linux pdf converter client server update ui design and error handling qubesos qubes app linux pdf converter client replace namedtuple with dataclass qubesos qubes app linux pdf converter client add initial version of tqdm output qubesos qubes app linux pdf converter client implement in place qubesos qubes app linux pdf converter client add archive dry run and in place qubesos qubes app linux pdf converter client server rename unsanitized file class qubesos qubes app linux pdf converter client server remove unused imports qubesos qubes app linux pdf converter server remove nonexistent pageerror qubesos qubes app linux pdf converter client server add representation file objects qubesos qubes app linux pdf converter client replace click echo qubesos qubes app linux pdf converter server update sending process qubesos qubes app linux pdf converter server update error handling qubesos qubes app linux pdf converter server remove recv pdf qubesos qubes app linux pdf converter client update error handling qubesos qubes app linux pdf converter client add cli parameter handling qubesos qubes app linux pdf converter client start batch processing code qubesos qubes app linux pdf converter rename source directory qubesos qubes app linux pdf converter server remove uneeded quoting in identify command qubesos qubes app linux pdf converter client remove unused recv qubesos qubes app linux pdf converter client remove unneeded pid prefix for tmpdir qubesos qubes app linux pdf converter client remove extra byte for send qubesos qubes app linux pdf converter client fix error handling for readexactly qubesos qubes app linux pdf converter meta remove wrapper qubesos qubes app linux pdf converter meta remove extra file qubesos qubes app linux pdf converter tests moved to pdf converter tests qubesos qubes app linux pdf converter meta remove extra file qubesos qubes app linux pdf converter client server move to pdf converter directory qubesos qubes app linux pdf converter client server update to asyncio and new model qubesos qubes app linux pdf converter client server update initial phase qubesos qubes app linux pdf converter client combine image dimension functions qubesos qubes app linux pdf converter client server update utility functions qubesos qubes app linux pdf converter client server add custom exceptions qubesos qubes app linux pdf converter client remove redundant check condition qubesos qubes app linux pdf converter client server replace recv with recvline b qubesos qubes app linux pdf converter update docs qubesos qubes app linux pdf converter add support for multiple files qubesos qubes app linux pdf converter readme update language qubesos qubes app linux pdf converter server client remove unecessary documentation qubesos qubes app linux pdf converter wrapper client server qualify variable names qubesos qubes app linux pdf converter wrapper add non file check qubesos qubes app linux pdf converter wrapper simplify argument parsing qubesos qubes app linux pdf converter server fix location of stdout flush call qubesos qubes app linux pdf converter client use og stdout file descriptor qubesos qubes app linux pdf converter server client grammar qubesos qubes app linux pdf converter server avoid errors if subprocess command fails qubesos qubes app linux pdf converter client handle eoferrors if server suddenly dies qubesos qubes app linux pdf converter client server fix stream flushing qubesos qubes app linux pdf converter wrapper move logging config to global namespace qubesos qubes app linux pdf converter wrapper trim down argument parsing qubesos qubes app linux pdf converter wrapper fix leftover debugging command qubesos qubes app linux pdf converter readme update usage example qubesos qubes app linux pdf converter add py extensions and update install commands qubesos qubes app linux pdf converter wrapper add documentation qubesos qubes app linux pdf converter server make check output return a string instead of a bytestring qubesos qubes app linux pdf converter client server add documentation qubesos qubes app linux pdf converter server refactor to python qubesos qubes app linux pdf converter client pep qubesos qubes app linux pdf converter client add qrexec wrapper functions qubesos qubes app linux pdf converter wrapper pep qubesos qubes app linux pdf converter wrapper client re add logging qubesos qubes app linux pdf converter client update to python qubesos qubes app linux pdf converter wrapper remove logging qubesos qubes app linux pdf converter wrapper remove unneeded main try block qubesos qubes app linux pdf converter wrapper prepare for multiple file support qubesos qubes app linux pdf converter wrapper add logging and trim options qubesos qubes app linux pdf converter wrapper update qvm convert pdf into python qubesos qubes app linux pdf converter readme remove extra parenthesis qubesos qubes app linux pdf converter merge remote tracking branch origin pr qubesos qubes app linux pdf converter debian conditional python version dependencies referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload app linux pdf converter current repo available days from now upload app linux pdf converter current dists repo you can choose subset of distributions like vm vm available days from now upload app linux pdf converter security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
1
77,058
9,978,309,287
IssuesEvent
2019-07-09 19:32:12
apollographql/apollo-server
https://api.github.com/repos/apollographql/apollo-server
closed
Error masking does not work
:memo: documentation :rose: has-reproduction 🚧👷‍♀️👷‍♂️🚧 in triage
<!--**Issue Labels** While not necessary, you can help organize our issues by labeling this issue when you open it. To add a label automatically, simply [x] mark the appropriate box below: - [x] has-reproduction - [ ] feature - [x] docs - [ ] blocking - [ ] good first issue To add a label not listed above, simply place `/label another-label-name` on a line by itself. --> An example given here https://github.com/apollographql/apollo-server/blob/ae9da10e625cf283568ba6d29cea8c3e69a7a03f/docs/source/features/errors.md#masking-and-logging-errors does not work returning empty object. I've reproduced this issue here https://github.com/almostprogrammer/format-error Also I'd like to know whether it's a good practice to handle errors in `formatError` avoiding duplicating code throughout the app?
1.0
Error masking does not work - <!--**Issue Labels** While not necessary, you can help organize our issues by labeling this issue when you open it. To add a label automatically, simply [x] mark the appropriate box below: - [x] has-reproduction - [ ] feature - [x] docs - [ ] blocking - [ ] good first issue To add a label not listed above, simply place `/label another-label-name` on a line by itself. --> An example given here https://github.com/apollographql/apollo-server/blob/ae9da10e625cf283568ba6d29cea8c3e69a7a03f/docs/source/features/errors.md#masking-and-logging-errors does not work returning empty object. I've reproduced this issue here https://github.com/almostprogrammer/format-error Also I'd like to know whether it's a good practice to handle errors in `formatError` avoiding duplicating code throughout the app?
non_test
error masking does not work issue labels while not necessary you can help organize our issues by labeling this issue when you open it to add a label automatically simply mark the appropriate box below has reproduction feature docs blocking good first issue to add a label not listed above simply place label another label name on a line by itself an example given here does not work returning empty object i ve reproduced this issue here also i d like to know whether it s a good practice to handle errors in formaterror avoiding duplicating code throughout the app
0
227,059
7,526,461,868
IssuesEvent
2018-04-13 14:06:20
webpack-contrib/sass-loader
https://api.github.com/repos/webpack-contrib/sass-loader
closed
Sass imports inside node_modules broken
priority: 1 (critical) semver: Patch severity: 2 (regression) type: Bug
It appears that 7.0 breaks sass modules we are using from npm. I guess it is related to https://github.com/webpack-contrib/sass-loader/pull/479 For example foundation-sites that we are using imports it's internal modules like this: ``` // Sass utilities @import 'util/util'; ``` Now, unfortunately 'util/util' is a valid file inside node_modules (from 'node-util' pkg), and the build fails with errror like this: ``` ERROR in ./src/style.scss (./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/style.scss) Module build failed: // Copyright Joyent, Inc. and other Node contributors. ^ Invalid CSS after "...N THE SOFTWARE.": expected 1 selector or at-rule, was "var formatRegExp = " in ###/webpack-sass-broken-imports/node_modules/util/util.js (line 1, column 1) @ ./src/style.scss 2:14-117 @ ./src/index.js ``` repro repo: https://github.com/larpo/webpack-sass-broken-module-imports Is there a way to revert to the old behaviour? I don't see any mentions on READMe about how to configure the aliasing behaviour.
1.0
Sass imports inside node_modules broken - It appears that 7.0 breaks sass modules we are using from npm. I guess it is related to https://github.com/webpack-contrib/sass-loader/pull/479 For example foundation-sites that we are using imports it's internal modules like this: ``` // Sass utilities @import 'util/util'; ``` Now, unfortunately 'util/util' is a valid file inside node_modules (from 'node-util' pkg), and the build fails with errror like this: ``` ERROR in ./src/style.scss (./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/style.scss) Module build failed: // Copyright Joyent, Inc. and other Node contributors. ^ Invalid CSS after "...N THE SOFTWARE.": expected 1 selector or at-rule, was "var formatRegExp = " in ###/webpack-sass-broken-imports/node_modules/util/util.js (line 1, column 1) @ ./src/style.scss 2:14-117 @ ./src/index.js ``` repro repo: https://github.com/larpo/webpack-sass-broken-module-imports Is there a way to revert to the old behaviour? I don't see any mentions on READMe about how to configure the aliasing behaviour.
non_test
sass imports inside node modules broken it appears that breaks sass modules we are using from npm i guess it is related to for example foundation sites that we are using imports it s internal modules like this sass utilities import util util now unfortunately util util is a valid file inside node modules from node util pkg and the build fails with errror like this error in src style scss node modules css loader node modules sass loader lib loader js src style scss module build failed copyright joyent inc and other node contributors invalid css after n the software expected selector or at rule was var formatregexp in webpack sass broken imports node modules util util js line column src style scss src index js repro repo is there a way to revert to the old behaviour i don t see any mentions on readme about how to configure the aliasing behaviour
0
12,826
15,211,211,283
IssuesEvent
2021-02-17 08:45:46
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Work flows are failed for participant manager and Participant manager datastore
Bug Process: Fixed
Please go through this links for the errors https://github.com/boston-tech/develop-fda-mystudies/actions/runs/558335875 https://github.com/boston-tech/develop-fda-mystudies/actions/runs/558335871
1.0
Work flows are failed for participant manager and Participant manager datastore - Please go through this links for the errors https://github.com/boston-tech/develop-fda-mystudies/actions/runs/558335875 https://github.com/boston-tech/develop-fda-mystudies/actions/runs/558335871
non_test
work flows are failed for participant manager and participant manager datastore please go through this links for the errors
0
534,040
15,608,018,495
IssuesEvent
2021-03-19 10:05:31
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
FileSystemAlreadyDefined Exception thrown when trying to load balos and projects from language server
Area/ProjectAPI Priority/Blocker SwanLakeDump Team/DevTools Type/Bug
**Description:** getting FileSystemAlreadyDefined exception when try to load balos within language server, following are the use-cases, 1. We need to read the package names within the distribution (standard libs and langlibs), for that we use the getPackages API in PackageRepository as follows ``` Environment environment = EnvironmentBuilder.buildDefault(); BallerinaDistribution ballerinaDistribution = BallerinaDistribution.from(environment); PackageRepository packageRepository = ballerinaDistribution.packageRepository(); Map<String, List<String>> pkgMap = packageRepository.getPackages(); ``` Above works without any issue 2. We read the balos and get semantic models ``` ProjectEnvironmentBuilder defaultBuilder = ProjectEnvironmentBuilder.getBuilder(environment); defaultBuilder.addCompilationCacheFactory(TempDirCompilationCache::from); BaloProject baloProject = BaloProject.loadProject(defaultBuilder, pkg.project().sourceRoot()); ``` the above also works fine and both environment instances are the same. Both of the use-cases are independent of use-case 3 3. Building a project, given the project root, ``` BuildProject.load(projectRoot, options); Use case three has the exception mentioned earlier. ``` Is there any way we can achieve the above three use-cases in the same process? **Steps to reproduce:** **Affected Versions:** SLP8 at least
1.0
FileSystemAlreadyDefined Exception thrown when trying to load balos and projects from language server - **Description:** getting FileSystemAlreadyDefined exception when try to load balos within language server, following are the use-cases, 1. We need to read the package names within the distribution (standard libs and langlibs), for that we use the getPackages API in PackageRepository as follows ``` Environment environment = EnvironmentBuilder.buildDefault(); BallerinaDistribution ballerinaDistribution = BallerinaDistribution.from(environment); PackageRepository packageRepository = ballerinaDistribution.packageRepository(); Map<String, List<String>> pkgMap = packageRepository.getPackages(); ``` Above works without any issue 2. We read the balos and get semantic models ``` ProjectEnvironmentBuilder defaultBuilder = ProjectEnvironmentBuilder.getBuilder(environment); defaultBuilder.addCompilationCacheFactory(TempDirCompilationCache::from); BaloProject baloProject = BaloProject.loadProject(defaultBuilder, pkg.project().sourceRoot()); ``` the above also works fine and both environment instances are the same. Both of the use-cases are independent of use-case 3 3. Building a project, given the project root, ``` BuildProject.load(projectRoot, options); Use case three has the exception mentioned earlier. ``` Is there any way we can achieve the above three use-cases in the same process? **Steps to reproduce:** **Affected Versions:** SLP8 at least
non_test
filesystemalreadydefined exception thrown when trying to load balos and projects from language server description getting filesystemalreadydefined exception when try to load balos within language server following are the use cases we need to read the package names within the distribution standard libs and langlibs for that we use the getpackages api in packagerepository as follows environment environment environmentbuilder builddefault ballerinadistribution ballerinadistribution ballerinadistribution from environment packagerepository packagerepository ballerinadistribution packagerepository map pkgmap packagerepository getpackages above works without any issue we read the balos and get semantic models projectenvironmentbuilder defaultbuilder projectenvironmentbuilder getbuilder environment defaultbuilder addcompilationcachefactory tempdircompilationcache from baloproject baloproject baloproject loadproject defaultbuilder pkg project sourceroot the above also works fine and both environment instances are the same both of the use cases are independent of use case building a project given the project root buildproject load projectroot options use case three has the exception mentioned earlier is there any way we can achieve the above three use cases in the same process steps to reproduce affected versions at least
0
341,340
30,581,966,678
IssuesEvent
2023-07-21 10:22:09
adoptium/infrastructure
https://api.github.com/repos/adoptium/infrastructure
closed
jdk11 java/beans/Beans/TypoInBeanDescription.java timeout on test-osuosl-aix715-ppc64-3
os:aix provider:osuosl testFail
java/beans/Beans/TypoInBeanDescription.java timeout on [test-osuosl-aix715-ppc64-3](https://ci.adoptopenjdk.net/computer/test-osuosl-aix715-ppc64-3) on passed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5044 - test suite/name? - Is there an existing issue elsewhere covering this? - Which machine(s) does it work on? [build-osuosl-aix71-ppc64-2](https://ci.adoptopenjdk.net/computer/build-osuosl-aix71-ppc64-2) - Which machine(s) does it fail on? [test-osuosl-aix715-ppc64-3](https://ci.adoptopenjdk.net/computer/test-osuosl-aix715-ppc64-3) - Do you have a link to a Grinder re-run if the test with the failure? rerun on failed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5042/ on passed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5044
1.0
jdk11 java/beans/Beans/TypoInBeanDescription.java timeout on test-osuosl-aix715-ppc64-3 - java/beans/Beans/TypoInBeanDescription.java timeout on [test-osuosl-aix715-ppc64-3](https://ci.adoptopenjdk.net/computer/test-osuosl-aix715-ppc64-3) on passed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5044 - test suite/name? - Is there an existing issue elsewhere covering this? - Which machine(s) does it work on? [build-osuosl-aix71-ppc64-2](https://ci.adoptopenjdk.net/computer/build-osuosl-aix71-ppc64-2) - Which machine(s) does it fail on? [test-osuosl-aix715-ppc64-3](https://ci.adoptopenjdk.net/computer/test-osuosl-aix715-ppc64-3) - Do you have a link to a Grinder re-run if the test with the failure? rerun on failed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5042/ on passed machine https://ci.adoptopenjdk.net/view/Test_grinder/job/Grinder/5044
test
java beans beans typoinbeandescription java timeout on test osuosl java beans beans typoinbeandescription java timeout on on passed machine test suite name is there an existing issue elsewhere covering this which machine s does it work on which machine s does it fail on do you have a link to a grinder re run if the test with the failure rerun on failed machine on passed machine
1
38,325
19,100,326,141
IssuesEvent
2021-11-29 21:38:56
golang/go
https://api.github.com/repos/golang/go
closed
cmd/compile: optimize comparisons to min/max (u)ints by checking for overflow
Performance help wanted NeedsFix
```go package p import "math" func f(x int64) bool { return x == math.MinInt64 } ``` On amd64, the core of this compiles to: ``` 0x0005 00005 (z.go:6) MOVQ $-9223372036854775808, CX 0x000f 00015 (z.go:6) CMPQ AX, CX 0x0012 00018 (z.go:6) SETEQ "".~r1+16(SP) ``` It would be cheaper and smaller instead to decrement CX and check the flags for underflow. A similar trick can be used for checking min and max ints and uints of all sizes. It might also be useful in the division fix-up code, where we must check for min int divisor. cc @randall77 @dr2chase @martisch @mundaym
True
cmd/compile: optimize comparisons to min/max (u)ints by checking for overflow - ```go package p import "math" func f(x int64) bool { return x == math.MinInt64 } ``` On amd64, the core of this compiles to: ``` 0x0005 00005 (z.go:6) MOVQ $-9223372036854775808, CX 0x000f 00015 (z.go:6) CMPQ AX, CX 0x0012 00018 (z.go:6) SETEQ "".~r1+16(SP) ``` It would be cheaper and smaller instead to decrement CX and check the flags for underflow. A similar trick can be used for checking min and max ints and uints of all sizes. It might also be useful in the division fix-up code, where we must check for min int divisor. cc @randall77 @dr2chase @martisch @mundaym
non_test
cmd compile optimize comparisons to min max u ints by checking for overflow go package p import math func f x bool return x math on the core of this compiles to z go movq cx z go cmpq ax cx z go seteq sp it would be cheaper and smaller instead to decrement cx and check the flags for underflow a similar trick can be used for checking min and max ints and uints of all sizes it might also be useful in the division fix up code where we must check for min int divisor cc martisch mundaym
0
313,986
26,967,856,076
IssuesEvent
2023-02-09 00:35:26
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
pull-kubernetes-unit failing
kind/failing-test
### Which jobs are failing? https://kubernetes.slack.com/archives/C7J9RP96G/p1675902134074689 https://prow.k8s.io/?job=pull-kubernetes-unit ### Which tests are failing? the entire job ### Since when has it been failing? recently ### Testgrid link _No response_ ### Reason for failure (if possible) prow issue, permission denied on /.gitconfig ### Anything else we need to know? https://github.com/kubernetes/test-infra/pull/28688#issuecomment-1423427848 ### Relevant SIG(s) /sig testing /assign
1.0
pull-kubernetes-unit failing - ### Which jobs are failing? https://kubernetes.slack.com/archives/C7J9RP96G/p1675902134074689 https://prow.k8s.io/?job=pull-kubernetes-unit ### Which tests are failing? the entire job ### Since when has it been failing? recently ### Testgrid link _No response_ ### Reason for failure (if possible) prow issue, permission denied on /.gitconfig ### Anything else we need to know? https://github.com/kubernetes/test-infra/pull/28688#issuecomment-1423427848 ### Relevant SIG(s) /sig testing /assign
test
pull kubernetes unit failing which jobs are failing which tests are failing the entire job since when has it been failing recently testgrid link no response reason for failure if possible prow issue permission denied on gitconfig anything else we need to know relevant sig s sig testing assign
1
79,199
7,698,323,423
IssuesEvent
2018-05-18 22:35:09
aspnet/SignalR
https://api.github.com/repos/aspnet/SignalR
closed
Flaky test: CanInvokeClientMethodFromServer
flaky test
This test [fails](http://aspnetci/viewLog.html?buildId=470227&buildTypeId=Lite_SignalR) occasionally with the following error: ``` System.AggregateException : One or more errors occurred. (Value cannot be null. Parameter name: path3) (The following constructor parameters did not have matching fixture data: ServerFixture`1 serverFixture) ---- System.ArgumentNullException : Value cannot be null. Parameter name: path3 ---- The following constructor parameters did not have matching fixture data: ServerFixture`1 serverFixture ----- Inner Stack Trace #1 (System.ArgumentNullException) ----- at System.IO.Path.Combine(String path1, String path2, String path3) at Microsoft.Extensions.Logging.Testing.AssemblyTestLog.Create(Assembly assembly, String baseDirectory) at Microsoft.Extensions.Logging.Testing.AssemblyTestLog.ForAssembly(Assembly assembly) at Microsoft.AspNetCore.SignalR.Tests.ServerFixture`1..ctor(ILoggerFactory loggerFactory) in /_/test/Microsoft.AspNetCore.SignalR.Tests.Utils/ServerFixture.cs:line 65 ----- Inner Stack Trace #2 (Xunit.Sdk.TestClassException) ----- ``` This happens on just regular builds of SignalR, so I assume it reproduces locally.
1.0
Flaky test: CanInvokeClientMethodFromServer - This test [fails](http://aspnetci/viewLog.html?buildId=470227&buildTypeId=Lite_SignalR) occasionally with the following error: ``` System.AggregateException : One or more errors occurred. (Value cannot be null. Parameter name: path3) (The following constructor parameters did not have matching fixture data: ServerFixture`1 serverFixture) ---- System.ArgumentNullException : Value cannot be null. Parameter name: path3 ---- The following constructor parameters did not have matching fixture data: ServerFixture`1 serverFixture ----- Inner Stack Trace #1 (System.ArgumentNullException) ----- at System.IO.Path.Combine(String path1, String path2, String path3) at Microsoft.Extensions.Logging.Testing.AssemblyTestLog.Create(Assembly assembly, String baseDirectory) at Microsoft.Extensions.Logging.Testing.AssemblyTestLog.ForAssembly(Assembly assembly) at Microsoft.AspNetCore.SignalR.Tests.ServerFixture`1..ctor(ILoggerFactory loggerFactory) in /_/test/Microsoft.AspNetCore.SignalR.Tests.Utils/ServerFixture.cs:line 65 ----- Inner Stack Trace #2 (Xunit.Sdk.TestClassException) ----- ``` This happens on just regular builds of SignalR, so I assume it reproduces locally.
test
flaky test caninvokeclientmethodfromserver this test occasionally with the following error system aggregateexception one or more errors occurred value cannot be null parameter name the following constructor parameters did not have matching fixture data serverfixture serverfixture system argumentnullexception value cannot be null parameter name the following constructor parameters did not have matching fixture data serverfixture serverfixture inner stack trace system argumentnullexception at system io path combine string string string at microsoft extensions logging testing assemblytestlog create assembly assembly string basedirectory at microsoft extensions logging testing assemblytestlog forassembly assembly assembly at microsoft aspnetcore signalr tests serverfixture ctor iloggerfactory loggerfactory in test microsoft aspnetcore signalr tests utils serverfixture cs line inner stack trace xunit sdk testclassexception this happens on just regular builds of signalr so i assume it reproduces locally
1
131,982
10,726,642,252
IssuesEvent
2019-10-28 09:51:44
wix/wix-style-react
https://api.github.com/repos/wix/wix-style-react
closed
Using enzyme as a dependency while it is served as a dev dependency
Priority:High Priority:Major testkit:bug
`wix-ui-test-utils` [`create-driver-factory`](https://github.com/wix/wix-ui/blob/7ad19b46dc5e73e36da1e2d0ac230709f78f5c2a/packages/wix-ui-test-utils/src/driver-factory/createDriverFactory.tsx#L4) is importing `enzyme`: We have drivers in wsr and in wix-ui-core which are importing this module from `wix-ui-test-utils` for example this [one](https://github.com/wix/wix-style-react/blob/5a24230f868a19bd65b2b3afffeb4950c7a7b2c9/src/Tooltip/Tooltip.driver.d.ts#L1) The outcome of it is - consumers which are using these kind of drivers without using installing `enzyme` themeselfs, their build will break. Priority: Major Solution: Remove enzyme import and [usage](https://github.com/wix/wix-ui/blob/7ad19b46dc5e73e36da1e2d0ac230709f78f5c2a/packages/wix-ui-test-utils/src/driver-factory/createDriverFactory.tsx#L15) which is just the `ReactWrapper` type **Update** Problem is that consumers are using `wix-ui-test-utils` directly and use `createDriverFactory` so it can break them.
1.0
Using enzyme as a dependency while it is served as a dev dependency - `wix-ui-test-utils` [`create-driver-factory`](https://github.com/wix/wix-ui/blob/7ad19b46dc5e73e36da1e2d0ac230709f78f5c2a/packages/wix-ui-test-utils/src/driver-factory/createDriverFactory.tsx#L4) is importing `enzyme`: We have drivers in wsr and in wix-ui-core which are importing this module from `wix-ui-test-utils` for example this [one](https://github.com/wix/wix-style-react/blob/5a24230f868a19bd65b2b3afffeb4950c7a7b2c9/src/Tooltip/Tooltip.driver.d.ts#L1) The outcome of it is - consumers which are using these kind of drivers without using installing `enzyme` themeselfs, their build will break. Priority: Major Solution: Remove enzyme import and [usage](https://github.com/wix/wix-ui/blob/7ad19b46dc5e73e36da1e2d0ac230709f78f5c2a/packages/wix-ui-test-utils/src/driver-factory/createDriverFactory.tsx#L15) which is just the `ReactWrapper` type **Update** Problem is that consumers are using `wix-ui-test-utils` directly and use `createDriverFactory` so it can break them.
test
using enzyme as a dependency while it is served as a dev dependency wix ui test utils is importing enzyme we have drivers in wsr and in wix ui core which are importing this module from wix ui test utils for example this the outcome of it is consumers which are using these kind of drivers without using installing enzyme themeselfs their build will break priority major solution remove enzyme import and which is just the reactwrapper type update problem is that consumers are using wix ui test utils directly and use createdriverfactory so it can break them
1
333,173
29,513,185,261
IssuesEvent
2023-06-04 07:00:55
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix MaskedArray.test_numpy_maskedarray_property_fill_value
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix MaskedArray.test_numpy_maskedarray_property_fill_value - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5033602665/jobs/9027925557" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix maskedarray test numpy maskedarray property fill value tensorflow img src torch img src numpy img src jax img src paddle img src
1
36,316
5,047,364,699
IssuesEvent
2016-12-20 09:10:29
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
MapLoaderMultiNodeTest.testLoadAgain_whenLoadedAllCalledMultipleTimes
Team: Core Type: Test-Failure
``` java.lang.AssertionError: expected:<20000> but was:<10000> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at org.junit.Assert.assertEquals(Assert.java:631) at com.hazelcast.map.mapstore.MapLoaderMultiNodeTest.testLoadAgain_whenLoadedAllCalledMultipleTimes(MapLoaderMultiNodeTest.java:127) ``` https://hazelcast-l337.ci.cloudbees.com/view/Hazelcast/job/Hazelcast-3.x-OracleJDK1.7/com.hazelcast$hazelcast/303/testReport/junit/com.hazelcast.map.mapstore/MapLoaderMultiNodeTest/testLoadAgain_whenLoadedAllCalledMultipleTimes/
1.0
MapLoaderMultiNodeTest.testLoadAgain_whenLoadedAllCalledMultipleTimes - ``` java.lang.AssertionError: expected:<20000> but was:<10000> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) at org.junit.Assert.assertEquals(Assert.java:631) at com.hazelcast.map.mapstore.MapLoaderMultiNodeTest.testLoadAgain_whenLoadedAllCalledMultipleTimes(MapLoaderMultiNodeTest.java:127) ``` https://hazelcast-l337.ci.cloudbees.com/view/Hazelcast/job/Hazelcast-3.x-OracleJDK1.7/com.hazelcast$hazelcast/303/testReport/junit/com.hazelcast.map.mapstore/MapLoaderMultiNodeTest/testLoadAgain_whenLoadedAllCalledMultipleTimes/
test
maploadermultinodetest testloadagain whenloadedallcalledmultipletimes java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at com hazelcast map mapstore maploadermultinodetest testloadagain whenloadedallcalledmultipletimes maploadermultinodetest java
1
41,646
12,836,314,066
IssuesEvent
2020-07-07 14:11:08
GNS3/gns3-web-ui
https://api.github.com/repos/GNS3/gns3-web-ui
closed
CVE-2019-14863 (Medium) detected in angular-1.3.14.min.js
security vulnerability
## CVE-2019-14863 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-1.3.14.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.3.14/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.3.14/angular.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/gns3-web-ui/node_modules/save-svg-as-png/viewboxtest.html</p> <p>Path to vulnerable library: /gns3-web-ui/node_modules/save-svg-as-png/viewboxtest.html</p> <p> Dependency Hierarchy: - :x: **angular-1.3.14.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GNS3/gns3-web-ui/commit/f0ec495ff41b611f048247d2919a7d1ebc5a2899">f0ec495ff41b611f048247d2919a7d1ebc5a2899</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a vulnerability in all angular versions before 1.5.0-beta.0, where after escaping the context of the web application, the web application delivers data to its users along with other trusted dynamic content, without validating it. <p>Publish Date: 2020-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14863>CVE-2019-14863</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/12524">https://github.com/angular/angular.js/pull/12524</a></p> <p>Release Date: 2020-01-02</p> <p>Fix Resolution: angular - v1.5.0-beta.1;org.webjars:angularjs:1.5.0-rc.0 </p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.3.14","isTransitiveDependency":false,"dependencyTree":"angular.js:1.3.14","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - v1.5.0-beta.1;org.webjars:angularjs:1.5.0-rc.0\t"}],"vulnerabilityIdentifier":"CVE-2019-14863","vulnerabilityDetails":"There is a vulnerability in all angular versions before 1.5.0-beta.0, where after escaping the context of the web application, the web application delivers data to its users along with other trusted dynamic content, without validating it.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14863","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-14863 (Medium) detected in angular-1.3.14.min.js - ## CVE-2019-14863 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-1.3.14.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.3.14/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.3.14/angular.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/gns3-web-ui/node_modules/save-svg-as-png/viewboxtest.html</p> <p>Path to vulnerable library: /gns3-web-ui/node_modules/save-svg-as-png/viewboxtest.html</p> <p> Dependency Hierarchy: - :x: **angular-1.3.14.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GNS3/gns3-web-ui/commit/f0ec495ff41b611f048247d2919a7d1ebc5a2899">f0ec495ff41b611f048247d2919a7d1ebc5a2899</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a vulnerability in all angular versions before 1.5.0-beta.0, where after escaping the context of the web application, the web application delivers data to its users along with other trusted dynamic content, without validating it. <p>Publish Date: 2020-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14863>CVE-2019-14863</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/12524">https://github.com/angular/angular.js/pull/12524</a></p> <p>Release Date: 2020-01-02</p> <p>Fix Resolution: angular - v1.5.0-beta.1;org.webjars:angularjs:1.5.0-rc.0 </p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.3.14","isTransitiveDependency":false,"dependencyTree":"angular.js:1.3.14","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - v1.5.0-beta.1;org.webjars:angularjs:1.5.0-rc.0\t"}],"vulnerabilityIdentifier":"CVE-2019-14863","vulnerabilityDetails":"There is a vulnerability in all angular versions before 1.5.0-beta.0, where after escaping the context of the web application, the web application delivers data to its users along with other trusted dynamic content, without validating it.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14863","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in angular min js cve medium severity vulnerability vulnerable library angular min js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to dependency file tmp ws scm web ui node modules save svg as png viewboxtest html path to vulnerable library web ui node modules save svg as png viewboxtest html dependency hierarchy x angular min js vulnerable library found in head commit a href vulnerability details there is a vulnerability in all angular versions before beta where after escaping the context of the web application the web application delivers data to its users along with other trusted dynamic content without validating it publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angular beta org webjars angularjs rc isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails there is a vulnerability in all angular versions before beta where after escaping the context of the web application the web application delivers data to its users along with other trusted dynamic content without validating it vulnerabilityurl
0
40,811
12,799,740,718
IssuesEvent
2020-07-02 15:52:10
mwilliams7197/zuul
https://api.github.com/repos/mwilliams7197/zuul
opened
WS-2019-0379 (Medium) detected in commons-codec-1.9.jar
security vulnerability
## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/zuul/zuul-sample/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - ribbon-httpclient-2.2.4.jar (Root Library) - httpclient-4.5.3.jar - :x: **commons-codec-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/zuul/commits/8588a970f404a14680bea9a4eda267023e612256">8588a970f404a14680bea9a4eda267023e612256</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.9","isTransitiveDependency":true,"dependencyTree":"com.netflix.ribbon:ribbon-httpclient:2.2.4;org.apache.httpcomponents:httpclient:4.5.3;commons-codec:commons-codec:1.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0379 (Medium) detected in commons-codec-1.9.jar - ## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/zuul/zuul-sample/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - ribbon-httpclient-2.2.4.jar (Root Library) - httpclient-4.5.3.jar - :x: **commons-codec-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/zuul/commits/8588a970f404a14680bea9a4eda267023e612256">8588a970f404a14680bea9a4eda267023e612256</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.9","isTransitiveDependency":true,"dependencyTree":"com.netflix.ribbon:ribbon-httpclient:2.2.4;org.apache.httpcomponents:httpclient:4.5.3;commons-codec:commons-codec:1.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file tmp ws scm zuul zuul sample build gradle path to vulnerable library home wss scanner gradle caches modules files commons codec commons codec commons codec jar home wss scanner gradle caches modules files commons codec commons codec commons codec jar dependency hierarchy ribbon httpclient jar root library httpclient jar x commons codec jar vulnerable library found in head commit a href vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation vulnerabilityurl
0
52,157
6,221,642,827
IssuesEvent
2017-07-10 06:40:06
arquillian/smart-testing
https://api.github.com/repos/arquillian/smart-testing
closed
Test Bed - functional tests improvement
Component: Test Bed Priority: High Status: In Progress Type: Feature
We need to have a better way for writing and running functional tests. Currently, we have modules `-ftest` which have little to no value. Ideally, we should provide an easy way to write tests which will illustrate real-world scenarios. So we would start with the clean project and then: * add smart-testing to it * define the priorities/strategies * make necessary changes for the given scenario * run the build and see what was the order of tests etc ### Suggested solution We could create a separate project to be used as a source for all our tests. Then we could apply all the changes as described above, execute the build using embedded maven and collect the results (either through grepping the log or looking at surefire reports). As for applying changes, for the moment I have two ideas in mind: * using templates / forge way * have already prepared changes as commits on some branch different than master (with tags) and just cherry-pick them. Suggestions welcome.
1.0
Test Bed - functional tests improvement - We need to have a better way for writing and running functional tests. Currently, we have modules `-ftest` which have little to no value. Ideally, we should provide an easy way to write tests which will illustrate real-world scenarios. So we would start with the clean project and then: * add smart-testing to it * define the priorities/strategies * make necessary changes for the given scenario * run the build and see what was the order of tests etc ### Suggested solution We could create a separate project to be used as a source for all our tests. Then we could apply all the changes as described above, execute the build using embedded maven and collect the results (either through grepping the log or looking at surefire reports). As for applying changes, for the moment I have two ideas in mind: * using templates / forge way * have already prepared changes as commits on some branch different than master (with tags) and just cherry-pick them. Suggestions welcome.
test
test bed functional tests improvement we need to have a better way for writing and running functional tests currently we have modules ftest which have little to no value ideally we should provide an easy way to write tests which will illustrate real world scenarios so we would start with the clean project and then add smart testing to it define the priorities strategies make necessary changes for the given scenario run the build and see what was the order of tests etc suggested solution we could create a separate project to be used as a source for all our tests then we could apply all the changes as described above execute the build using embedded maven and collect the results either through grepping the log or looking at surefire reports as for applying changes for the moment i have two ideas in mind using templates forge way have already prepared changes as commits on some branch different than master with tags and just cherry pick them suggestions welcome
1
331,634
29,044,885,594
IssuesEvent
2023-05-13 12:32:25
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts - cases security and spaces enabled: trial push_case memoryless server should push to a connector without mapping
failed-test Team:ResponseOps
A test failed on a tracked branch ``` Error: expected 200 "OK", got 403 "Forbidden" at Test._assertStatus (node_modules/supertest/lib/test.js:268:12) at Test._assertFunction (node_modules/supertest/lib/test.js:283:11) at Test.assert (node_modules/supertest/lib/test.js:173:18) at localAssert (node_modules/supertest/lib/test.js:131:12) at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/supertest/lib/test.js:128:5 at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3) at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/superagent/lib/node/index.js:916:18 at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7) at IncomingMessage.emit (node:events:525:35) at endReadableNT (node:internal/streams/readable:1358:12) at processTicksAndRejections (node:internal/process/task_queues:83:21) ``` First failure: [CI Build - 8.8](https://buildkite.com/elastic/kibana-on-merge/builds/30378#01881150-c3e6-4d3d-8eb0-32da324fbeb3) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts","test.name":"cases security and spaces enabled: trial push_case memoryless server should push to a connector without mapping","test.failCount":2}} -->
1.0
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts - cases security and spaces enabled: trial push_case memoryless server should push to a connector without mapping - A test failed on a tracked branch ``` Error: expected 200 "OK", got 403 "Forbidden" at Test._assertStatus (node_modules/supertest/lib/test.js:268:12) at Test._assertFunction (node_modules/supertest/lib/test.js:283:11) at Test.assert (node_modules/supertest/lib/test.js:173:18) at localAssert (node_modules/supertest/lib/test.js:131:12) at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/supertest/lib/test.js:128:5 at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3) at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/superagent/lib/node/index.js:916:18 at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7) at IncomingMessage.emit (node:events:525:35) at endReadableNT (node:internal/streams/readable:1358:12) at processTicksAndRejections (node:internal/process/task_queues:83:21) ``` First failure: [CI Build - 8.8](https://buildkite.com/elastic/kibana-on-merge/builds/30378#01881150-c3e6-4d3d-8eb0-32da324fbeb3) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts","test.name":"cases security and spaces enabled: trial push_case memoryless server should push to a connector without mapping","test.failCount":2}} -->
test
failing test x pack case api integration tests x pack test cases api integration security and spaces tests trial cases push case·ts cases security and spaces enabled trial push case memoryless server should push to a connector without mapping a test failed on a tracked branch error expected ok got forbidden at test assertstatus node modules supertest lib test js at test assertfunction node modules supertest lib test js at test assert node modules supertest lib test js at localassert node modules supertest lib test js at home buildkite agent builds kb spot elastic kibana on merge kibana node modules supertest lib test js at test request callback node modules superagent lib node index js at home buildkite agent builds kb spot elastic kibana on merge kibana node modules superagent lib node index js at incomingmessage node modules superagent lib node parsers json js at incomingmessage emit node events at endreadablent node internal streams readable at processticksandrejections node internal process task queues first failure
1
86,251
8,030,103,707
IssuesEvent
2018-07-27 18:22:30
italia/spid-testenv2
https://api.github.com/repos/italia/spid-testenv2
closed
AuthnRequest: errore "KeyError: 'binding'"
bug needs regression test
Inviando questa AuthnRequest: ```xml <saml2p:AuthnRequest ForceAuthn="false" IssueInstant="2018-07-24T17:15:47Z" Version="2.0" Destination="http://localhost:8088/sso" ID="_b09d06c33ba8649e14b34421d5a5d102" AttributeConsumingServiceIndex="1" AssertionConsumerServiceIndex="0" xmlns:saml2p="urn:oasis:names:tc:SAML:2.0:protocol" xmlns:saml2="urn:oasis:names:tc:SAML:2.0:assertion"><saml2:Issuer Format="urn:oasis:names:tc:SAML:2.0:nameid-format:entity" NameQualifier="https://www.prova.it/">https://www.prova.it/</saml2:Issuer><saml2p:NameIDPolicy Format="urn:oasis:names:tc:SAML:2.0:nameid-format:transient"></saml2p:NameIDPolicy><saml2p:RequestedAuthnContext Comparison="minimum"><saml2:AuthnContextClassRef>https://www.spid.gov.it/SpidL1</saml2:AuthnContextClassRef></saml2p:RequestedAuthnContext></saml2p:AuthnRequest> ``` ottengo il seguente errore: ``` Traceback (most recent call last): File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1741, in handle_exception reraise(exc_type, exc_value, tb) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/_compat.py", line 35, in reraise raise value File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/_compat.py", line 35, in reraise raise value File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/Users/alranel/Documents/git/TD/spid-testenv2/spid-testenv.py", line 1020, in single_sign_on_service atcss = self.server.metadata.attribute_consuming_service(sp_id) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 1108, in attribute_consuming_service "attribute_consuming_service", binding) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 978, in service srvs = _md.service(entity_id, typ, service, binding) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 575, in service if srv["binding"] == binding: KeyError: 'binding' ```
1.0
AuthnRequest: errore "KeyError: 'binding'" - Inviando questa AuthnRequest: ```xml <saml2p:AuthnRequest ForceAuthn="false" IssueInstant="2018-07-24T17:15:47Z" Version="2.0" Destination="http://localhost:8088/sso" ID="_b09d06c33ba8649e14b34421d5a5d102" AttributeConsumingServiceIndex="1" AssertionConsumerServiceIndex="0" xmlns:saml2p="urn:oasis:names:tc:SAML:2.0:protocol" xmlns:saml2="urn:oasis:names:tc:SAML:2.0:assertion"><saml2:Issuer Format="urn:oasis:names:tc:SAML:2.0:nameid-format:entity" NameQualifier="https://www.prova.it/">https://www.prova.it/</saml2:Issuer><saml2p:NameIDPolicy Format="urn:oasis:names:tc:SAML:2.0:nameid-format:transient"></saml2p:NameIDPolicy><saml2p:RequestedAuthnContext Comparison="minimum"><saml2:AuthnContextClassRef>https://www.spid.gov.it/SpidL1</saml2:AuthnContextClassRef></saml2p:RequestedAuthnContext></saml2p:AuthnRequest> ``` ottengo il seguente errore: ``` Traceback (most recent call last): File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1741, in handle_exception reraise(exc_type, exc_value, tb) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/_compat.py", line 35, in reraise raise value File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/_compat.py", line 35, in reraise raise value File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/Users/alranel/Documents/git/TD/spid-testenv2/spid-testenv.py", line 1020, in single_sign_on_service atcss = self.server.metadata.attribute_consuming_service(sp_id) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 1108, in attribute_consuming_service "attribute_consuming_service", binding) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 978, in service srvs = _md.service(entity_id, typ, service, binding) File "/Users/alranel/Documents/git/TD/spid-testenv2/env/lib/python3.7/site-packages/saml2/mdstore.py", line 575, in service if srv["binding"] == binding: KeyError: 'binding' ```
test
authnrequest errore keyerror binding inviando questa authnrequest xml ottengo il seguente errore traceback most recent call last file users alranel documents git td spid env lib site packages flask app py line in call return self wsgi app environ start response file users alranel documents git td spid env lib site packages flask app py line in wsgi app response self handle exception e file users alranel documents git td spid env lib site packages flask app py line in handle exception reraise exc type exc value tb file users alranel documents git td spid env lib site packages flask compat py line in reraise raise value file users alranel documents git td spid env lib site packages flask app py line in wsgi app response self full dispatch request file users alranel documents git td spid env lib site packages flask app py line in full dispatch request rv self handle user exception e file users alranel documents git td spid env lib site packages flask app py line in handle user exception reraise exc type exc value tb file users alranel documents git td spid env lib site packages flask compat py line in reraise raise value file users alranel documents git td spid env lib site packages flask app py line in full dispatch request rv self dispatch request file users alranel documents git td spid env lib site packages flask app py line in dispatch request return self view functions req view args file users alranel documents git td spid spid testenv py line in single sign on service atcss self server metadata attribute consuming service sp id file users alranel documents git td spid env lib site packages mdstore py line in attribute consuming service attribute consuming service binding file users alranel documents git td spid env lib site packages mdstore py line in service srvs md service entity id typ service binding file users alranel documents git td spid env lib site packages mdstore py line in service if srv binding keyerror binding
1
144,824
22,573,395,088
IssuesEvent
2022-06-28 04:02:31
szwathub/LeetCode.swift
https://api.github.com/repos/szwathub/LeetCode.swift
closed
715. Range 模块
question: hard design ordered set segment tree
# [715. Range 模块](https://leetcode.cn/problems/range-module/) Range模块是跟踪数字范围的模块。设计一个数据结构来跟踪表示为 **半开区间** 的范围并查询它们。 **半开区间** `[left, right)` 表示所有 `left <= x < right` 的实数 `x` 。 实现 **RangeModule** 类: - `RangeModule()` 初始化数据结构的对象。 - `void addRange(int left, int right)` 添加 **半开区间** `[left, right)`,跟踪该区间中的每个实数。添加与当前跟踪的数字部分重叠的区间时,应当添加在区间 `[left, right)` 中尚未跟踪的任何数字到该区间中。 - `boolean queryRange(int left, int right)` 只有在当前正在跟踪区间 **[left, right)** 中的每一个实数时,才返回 `true` ,否则返回 `false` 。 - `void removeRange(int left, int right)` 停止跟踪 **半开区间** `[left, right)` 中当前正在跟踪的每个实数。 **示例 1:** ``` 输入 ["RangeModule", "addRange", "removeRange", "queryRange", "queryRange", "queryRange"] [[], [10, 20], [14, 16], [10, 14], [13, 15], [16, 17]] 输出 [null, null, null, true, false, true] 解释 RangeModule rangeModule = new RangeModule(); rangeModule.addRange(10, 20); rangeModule.removeRange(14, 16); rangeModule.queryRange(10, 14); 返回 true (区间 [10, 14) 中的每个数都正在被跟踪) rangeModule.queryRange(13, 15); 返回 false(未跟踪区间 [13, 15) 中像 14, 14.03, 14.17 这样的数字) rangeModule.queryRange(16, 17); 返回 true (尽管执行了删除操作,区间 [16, 17) 中的数字 16 仍然会被跟踪) ``` **提示:** - 1 <= left < right <= 10<sup>9</sup> - 在单个测试用例中,对 `addRange` 、  `queryRange` 和 `removeRange` 的调用总数不超过 10<sup>4</sup> 次 ---- 来源:力扣(LeetCode) 链接:https://leetcode.cn/problems/range-module
1.0
715. Range 模块 - # [715. Range 模块](https://leetcode.cn/problems/range-module/) Range模块是跟踪数字范围的模块。设计一个数据结构来跟踪表示为 **半开区间** 的范围并查询它们。 **半开区间** `[left, right)` 表示所有 `left <= x < right` 的实数 `x` 。 实现 **RangeModule** 类: - `RangeModule()` 初始化数据结构的对象。 - `void addRange(int left, int right)` 添加 **半开区间** `[left, right)`,跟踪该区间中的每个实数。添加与当前跟踪的数字部分重叠的区间时,应当添加在区间 `[left, right)` 中尚未跟踪的任何数字到该区间中。 - `boolean queryRange(int left, int right)` 只有在当前正在跟踪区间 **[left, right)** 中的每一个实数时,才返回 `true` ,否则返回 `false` 。 - `void removeRange(int left, int right)` 停止跟踪 **半开区间** `[left, right)` 中当前正在跟踪的每个实数。 **示例 1:** ``` 输入 ["RangeModule", "addRange", "removeRange", "queryRange", "queryRange", "queryRange"] [[], [10, 20], [14, 16], [10, 14], [13, 15], [16, 17]] 输出 [null, null, null, true, false, true] 解释 RangeModule rangeModule = new RangeModule(); rangeModule.addRange(10, 20); rangeModule.removeRange(14, 16); rangeModule.queryRange(10, 14); 返回 true (区间 [10, 14) 中的每个数都正在被跟踪) rangeModule.queryRange(13, 15); 返回 false(未跟踪区间 [13, 15) 中像 14, 14.03, 14.17 这样的数字) rangeModule.queryRange(16, 17); 返回 true (尽管执行了删除操作,区间 [16, 17) 中的数字 16 仍然会被跟踪) ``` **提示:** - 1 <= left < right <= 10<sup>9</sup> - 在单个测试用例中,对 `addRange` 、  `queryRange` 和 `removeRange` 的调用总数不超过 10<sup>4</sup> 次 ---- 来源:力扣(LeetCode) 链接:https://leetcode.cn/problems/range-module
non_test
range 模块 range模块是跟踪数字范围的模块。设计一个数据结构来跟踪表示为 半开区间 的范围并查询它们。 半开区间   left right  表示所有  left x right  的实数 x 。 实现 rangemodule 类 rangemodule  初始化数据结构的对象。 void addrange int left int right 添加 半开区间   left right ,跟踪该区间中的每个实数。添加与当前跟踪的数字部分重叠的区间时,应当添加在区间  left right  中尚未跟踪的任何数字到该区间中。 boolean queryrange int left int right  只有在当前正在跟踪区间  left right  中的每一个实数时,才返回 true  ,否则返回 false 。 void removerange int left int right  停止跟踪 半开区间   left right  中当前正在跟踪的每个实数。 示例 : 输入 输出 解释 rangemodule rangemodule new rangemodule rangemodule addrange rangemodule removerange rangemodule queryrange 返回 true (区间 中的每个数都正在被跟踪) rangemodule queryrange 返回 false(未跟踪区间 中像 这样的数字) rangemodule queryrange 返回 true (尽管执行了删除操作,区间 中的数字 仍然会被跟踪) 提示: 在单个测试用例中,对  addrange  、  queryrange  和 removerange 的调用总数不超过   次 来源:力扣(leetcode) 链接:
0
292,431
25,213,728,158
IssuesEvent
2022-11-14 07:19:12
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
opened
Release 4.3.10
release test/4.3.10
Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.10 release. Close the issue only after the GA release. ## Release info | | | |---------------------------------------|------------------------------------------------------| | Release name| Wazuh v4.3.10| | Tag | https://github.com/wazuh/wazuh/tree/v4.3.10 | | Branch | https://github.com/wazuh/wazuh/tree/4.3 | | Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md | | Latest release candidate (GA version) | N/A | | ## Release candidates As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version. | Version | Issue | Tag | Status | |---------|-------|-------------------------------------------------|-----------| | RC1 | - | [v4.3.10-rc1](https://github.com/wazuh/wazuh/tree/v4.3.10-rc1) | In progress | - _In progress: Fixes are being implemented._ - _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._ - _Closed: Fixes and QA process has finished._ ## Tests All the test results for this release can be found in the [v4.3.10 project](https://github.com/orgs/wazuh/projects/23/views/2). ## Fixes/Known issues All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.10 project](https://github.com/orgs/wazuh/projects/23/views/1).
1.0
Release 4.3.10 - Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.10 release. Close the issue only after the GA release. ## Release info | | | |---------------------------------------|------------------------------------------------------| | Release name| Wazuh v4.3.10| | Tag | https://github.com/wazuh/wazuh/tree/v4.3.10 | | Branch | https://github.com/wazuh/wazuh/tree/4.3 | | Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md | | Latest release candidate (GA version) | N/A | | ## Release candidates As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version. | Version | Issue | Tag | Status | |---------|-------|-------------------------------------------------|-----------| | RC1 | - | [v4.3.10-rc1](https://github.com/wazuh/wazuh/tree/v4.3.10-rc1) | In progress | - _In progress: Fixes are being implemented._ - _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._ - _Closed: Fixes and QA process has finished._ ## Tests All the test results for this release can be found in the [v4.3.10 project](https://github.com/orgs/wazuh/projects/23/views/2). ## Fixes/Known issues All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.10 project](https://github.com/orgs/wazuh/projects/23/views/1).
test
release main issue gathering all the tasks regarding the preparation and validation of wazuh version release close the issue only after the ga release release info release name wazuh tag branch changelog latest release candidate ga version n a release candidates as usual in order to publish one or more release candidate versions are generated and fully tested in the links below you can check the status and evolution of each version version issue tag status in progress in progress fixes are being implemented in review all the fixes have been implemented the rc is now tagged and the qa process kicks off closed fixes and qa process has finished tests all the test results for this release can be found in the fixes known issues all the fixes and the issues detected in the testing phase that won t be solved in this version can be found in the
1
320,723
27,454,617,429
IssuesEvent
2023-03-02 20:13:13
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Orçamento - Todas - Brumadinho
generalization test development tag - Orçamento subtag - Todas
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Todas para o Município de Brumadinho.
1.0
Teste de generalizacao para a tag Orçamento - Todas - Brumadinho - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Todas para o Município de Brumadinho.
test
teste de generalizacao para a tag orçamento todas brumadinho dod realizar o teste de generalização do validador da tag orçamento todas para o município de brumadinho
1
434,918
30,476,248,543
IssuesEvent
2023-07-17 16:43:36
labring/sealos
https://api.github.com/repos/labring/sealos
closed
doc: sealos load and sealos save doc for new release
kind/documentation
When this version is released, the corresponding save and load documents need to be updated _Originally posted by @cuisongliu in https://github.com/labring/sealos/issues/3442#issuecomment-1614194081_
1.0
doc: sealos load and sealos save doc for new release - When this version is released, the corresponding save and load documents need to be updated _Originally posted by @cuisongliu in https://github.com/labring/sealos/issues/3442#issuecomment-1614194081_
non_test
doc sealos load and sealos save doc for new release when this version is released the corresponding save and load documents need to be updated originally posted by cuisongliu in
0
231,371
18,763,056,518
IssuesEvent
2021-11-05 18:59:32
astropy/astropy
https://api.github.com/repos/astropy/astropy
closed
[5.0RC1] exact double comparison in test_spline.py, exposed by Debian/ppc64el
testing Bug modeling
### Description Some tests in `astropy/modeling/tests/test_spline.py` compare doubles by equal, which not always works. Specifically on the PowerPC 64 Little Endian architecture on Debian, this gives failures ### Expected behavior Tests should pass :-) ### Actual behavior ``` _________________ TestSpline1D.test_interpolate_fitter[2-None] _________________ self = <astropy.modeling.tests.test_spline.TestSpline1D object at 0x7ffeb53a7a90> w = None, k = 2 @pytest.mark.parametrize('w', wieght_tests) @pytest.mark.parametrize('k', degree_tests) def test_interpolate_fitter(self, w, k): [...] assert (fit_spl.t == spline._eval_args[0]).all() assert (fit_spl.c == spline._eval_args[1]).all() assert (fitter.fit_info['spline']._eval_args[0] == spline._eval_args[0]).all() assert (fitter.fit_info['spline']._eval_args[1] == spline._eval_args[1]).all() assert spline.get_residual() == fitter.fit_info['resid'] == 0 > assert (fit_spl(self.x) == spline(self.x)).all() E assert False E + where False = <built-in method all of numpy.ndarray object at 0x7ffeb52a9030>() E + where <built-in method all of numpy.ndarray object at 0x7ffeb52a9030> = array([ 0.073... 0.05145015]) == array([ 0.073... 0.05145015]) E Use -v to get the full diff.all ``` The same happens on `test_smoothing_fitter` (line 1207) and `test_exact_knots_fitter` (line 1267), all fails several times for different parameters. [Full build log](https://buildd.debian.org/status/fetch.php?pkg=astropy&arch=ppc64el&ver=5.0%7Erc1-1&stamp=1636027474&raw=0) ### Steps to Reproduce This happens when running on PowerPC 64 bit little endian. ### System Details platform linux -- Python 3.9.7, pytest-6.2.5, py-1.10.0, pluggy-0.13.0 Matplotlib: 3.3.4 Freetype: 2.11.0 Running tests with Astropy version 5.0-rc1. Running tests in .pybuild/cpython3_3.9/build. Date: 2021-11-04T11:57:20 Platform: Linux-5.10.0-9-powerpc64le-ppc64le-with-glibc2.32 Executable: /usr/bin/python3.9 Full Python Version: 3.9.7 (default, Sep 24 2021, 09:43:00) [GCC 10.3.0] encodings: sys: utf-8, locale: UTF-8, filesystem: utf-8 byteorder: little float info: dig: 15, mant_dig: 15 Package versions: Numpy: 1.19.5 Scipy: 1.7.1 Matplotlib: 3.3.4 h5py: 3.3.0 Pandas: not available PyERFA: 2.0.0 Cython: 0.29.24 Scikit-image: not available asdf: not available pyarrow: not available Using Astropy options: remote_data: none. ARCH_ON_CI: undefined IS_CRON: undefined rootdir: /<<PKGBUILDDIR>>, configfile: setup.cfg plugins: mpl-0.11, remotedata-0.3.2, openfiles-0.5.0, mock-3.6.1, filter-subpackage-0.1.1, doctestplus-0.11.0, cov-3.0.0, astropy-header-0.1.2, arraydiff-0.3, hypothesis-5.43.3
1.0
[5.0RC1] exact double comparison in test_spline.py, exposed by Debian/ppc64el - ### Description Some tests in `astropy/modeling/tests/test_spline.py` compare doubles by equal, which not always works. Specifically on the PowerPC 64 Little Endian architecture on Debian, this gives failures ### Expected behavior Tests should pass :-) ### Actual behavior ``` _________________ TestSpline1D.test_interpolate_fitter[2-None] _________________ self = <astropy.modeling.tests.test_spline.TestSpline1D object at 0x7ffeb53a7a90> w = None, k = 2 @pytest.mark.parametrize('w', wieght_tests) @pytest.mark.parametrize('k', degree_tests) def test_interpolate_fitter(self, w, k): [...] assert (fit_spl.t == spline._eval_args[0]).all() assert (fit_spl.c == spline._eval_args[1]).all() assert (fitter.fit_info['spline']._eval_args[0] == spline._eval_args[0]).all() assert (fitter.fit_info['spline']._eval_args[1] == spline._eval_args[1]).all() assert spline.get_residual() == fitter.fit_info['resid'] == 0 > assert (fit_spl(self.x) == spline(self.x)).all() E assert False E + where False = <built-in method all of numpy.ndarray object at 0x7ffeb52a9030>() E + where <built-in method all of numpy.ndarray object at 0x7ffeb52a9030> = array([ 0.073... 0.05145015]) == array([ 0.073... 0.05145015]) E Use -v to get the full diff.all ``` The same happens on `test_smoothing_fitter` (line 1207) and `test_exact_knots_fitter` (line 1267), all fails several times for different parameters. [Full build log](https://buildd.debian.org/status/fetch.php?pkg=astropy&arch=ppc64el&ver=5.0%7Erc1-1&stamp=1636027474&raw=0) ### Steps to Reproduce This happens when running on PowerPC 64 bit little endian. ### System Details platform linux -- Python 3.9.7, pytest-6.2.5, py-1.10.0, pluggy-0.13.0 Matplotlib: 3.3.4 Freetype: 2.11.0 Running tests with Astropy version 5.0-rc1. Running tests in .pybuild/cpython3_3.9/build. Date: 2021-11-04T11:57:20 Platform: Linux-5.10.0-9-powerpc64le-ppc64le-with-glibc2.32 Executable: /usr/bin/python3.9 Full Python Version: 3.9.7 (default, Sep 24 2021, 09:43:00) [GCC 10.3.0] encodings: sys: utf-8, locale: UTF-8, filesystem: utf-8 byteorder: little float info: dig: 15, mant_dig: 15 Package versions: Numpy: 1.19.5 Scipy: 1.7.1 Matplotlib: 3.3.4 h5py: 3.3.0 Pandas: not available PyERFA: 2.0.0 Cython: 0.29.24 Scikit-image: not available asdf: not available pyarrow: not available Using Astropy options: remote_data: none. ARCH_ON_CI: undefined IS_CRON: undefined rootdir: /<<PKGBUILDDIR>>, configfile: setup.cfg plugins: mpl-0.11, remotedata-0.3.2, openfiles-0.5.0, mock-3.6.1, filter-subpackage-0.1.1, doctestplus-0.11.0, cov-3.0.0, astropy-header-0.1.2, arraydiff-0.3, hypothesis-5.43.3
test
exact double comparison in test spline py exposed by debian description some tests in astropy modeling tests test spline py compare doubles by equal which not always works specifically on the powerpc little endian architecture on debian this gives failures expected behavior tests should pass actual behavior test interpolate fitter self w none k pytest mark parametrize w wieght tests pytest mark parametrize k degree tests def test interpolate fitter self w k assert fit spl t spline eval args all assert fit spl c spline eval args all assert fitter fit info eval args spline eval args all assert fitter fit info eval args spline eval args all assert spline get residual fitter fit info assert fit spl self x spline self x all e assert false e where false e where array array e use v to get the full diff all the same happens on test smoothing fitter line and test exact knots fitter line all fails several times for different parameters steps to reproduce this happens when running on powerpc bit little endian system details platform linux python pytest py pluggy matplotlib freetype running tests with astropy version running tests in pybuild build date platform linux with executable usr bin full python version default sep encodings sys utf locale utf filesystem utf byteorder little float info dig mant dig package versions numpy scipy matplotlib pandas not available pyerfa cython scikit image not available asdf not available pyarrow not available using astropy options remote data none arch on ci undefined is cron undefined rootdir configfile setup cfg plugins mpl remotedata openfiles mock filter subpackage doctestplus cov astropy header arraydiff hypothesis
1
344,992
30,779,821,130
IssuesEvent
2023-07-31 09:14:46
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.5.0 - Alpha 1 - Workload benchmarks metrics
type/test level/subtask
The following issue aims to run all `workload benchmarks` for the current release candidate, report the results, and open new issues for any encountered errors. ## Workload benchmarks metrics information | | | |-----------------------------------------------|--------------------------------------------| | **Main release candidate issue** |https://github.com/wazuh/wazuh/issues/18058| | **Version** |4.5.0| | **Release candidate #** |Alpha 1| | **Tag** |[v4.5.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.5.0-alpha1)| | **Previous Workload benchmarks metrics issue**|https://github.com/wazuh/wazuh/issues/16975| ## Test configuration All tests will be run and workload performance metrics will be obtained for the following clustered environment configurations: | | | |-------------------|--------------------| | **# Agents** | **# Worker nodes** | |50000|25| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions All tests have been executed and the results can be found [here](https://github.com/wazuh/wazuh/issues/18113#issuecomment-1657894781). All tests have passed and the fails have been reported or justified. I therefore conclude that this issue is finished and OK for this release candidate. ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @Selutario
1.0
Release 4.5.0 - Alpha 1 - Workload benchmarks metrics - The following issue aims to run all `workload benchmarks` for the current release candidate, report the results, and open new issues for any encountered errors. ## Workload benchmarks metrics information | | | |-----------------------------------------------|--------------------------------------------| | **Main release candidate issue** |https://github.com/wazuh/wazuh/issues/18058| | **Version** |4.5.0| | **Release candidate #** |Alpha 1| | **Tag** |[v4.5.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.5.0-alpha1)| | **Previous Workload benchmarks metrics issue**|https://github.com/wazuh/wazuh/issues/16975| ## Test configuration All tests will be run and workload performance metrics will be obtained for the following clustered environment configurations: | | | |-------------------|--------------------| | **# Agents** | **# Worker nodes** | |50000|25| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions All tests have been executed and the results can be found [here](https://github.com/wazuh/wazuh/issues/18113#issuecomment-1657894781). All tests have passed and the fails have been reported or justified. I therefore conclude that this issue is finished and OK for this release candidate. ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @Selutario
test
release alpha workload benchmarks metrics the following issue aims to run all workload benchmarks for the current release candidate report the results and open new issues for any encountered errors workload benchmarks metrics information main release candidate issue version release candidate alpha tag previous workload benchmarks metrics issue test configuration all tests will be run and workload performance metrics will be obtained for the following clustered environment configurations agents worker nodes test report procedure all individual test checks must be marked as pass the test ran successfully xfail the test was expected to fail and it failed it must be properly justified and reported in an issue skip the test was not run it must be properly justified and reported in an issue fail the test failed a new issue must be opened to evaluate and address the problem all test results must have one the following statuses green circle all checks passed red circle there is at least one failed check yellow circle there is at least one expected fail or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause it must be included in the fixes section of the current release candidate main issue any expected fail or skipped test must have an issue justifying the reason all auditors must validate the justification for an expected fail or skipped test an extended report of the test results must be attached as a zip or txt this report can be used by the auditors to dig deeper into any possible failures and details conclusions all tests have been executed and the results can be found all tests have passed and the fails have been reported or justified i therefore conclude that this issue is finished and ok for this release candidate auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue selutario
1
268,472
8,407,193,573
IssuesEvent
2018-10-11 20:11:54
clearlinux/swupd-client
https://api.github.com/repos/clearlinux/swupd-client
opened
Testlib: Consolidate Minversion Implementation
low priority
The update_minversion and bump_format testlib functions both perform a minversion update, but with separate implementations. These implementations should be consolidated.
1.0
Testlib: Consolidate Minversion Implementation - The update_minversion and bump_format testlib functions both perform a minversion update, but with separate implementations. These implementations should be consolidated.
non_test
testlib consolidate minversion implementation the update minversion and bump format testlib functions both perform a minversion update but with separate implementations these implementations should be consolidated
0
88,709
17,647,679,761
IssuesEvent
2021-08-20 08:43:27
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Code editor not accepting block element (<h3>) inside anchor tag (<a>)
[Type] Bug [Feature] Code Editor [Package] Autop
**Describe the bug** Whenever editing a page/post through the code editor instead of the visual one and adding a block element inside an anchor tag, whenever focusing out or saving the page, the editor automatically changes the code creating an orphan anchor tag. **To reproduce** Steps to reproduce the behaviour: 1. Go to 'edit post or page' 2. Click on '3-dotted menu > code editor` 3. Paste the mark up below 4. Click outside the editor (focusing out) 5. The markup changes automatically 6. Publish it 7. The mark up on front end changes even further ```html <ul> <li> <a href='#'> <h3>Title</h3> </a> </li> </ul> ``` **Expected behaviour** Since the markup added is valid HTML I would expect it to be kept as is. But instead, on focusing out the editor, it is changed to (adjusted indentation for readability) ```html <ul> <li> <a href="#"> <h3>Title</h3> </a> <a href="#"></a> </li> </ul> ``` However, when publishing, it changes the markup even further to: ```html <ul> <li> <a href="#"> </p> <h3>Title</h3> <p> </a> <a href="#"></a> </li> </ul> ``` **Additional context** - WordPress 5.3
1.0
Code editor not accepting block element (<h3>) inside anchor tag (<a>) - **Describe the bug** Whenever editing a page/post through the code editor instead of the visual one and adding a block element inside an anchor tag, whenever focusing out or saving the page, the editor automatically changes the code creating an orphan anchor tag. **To reproduce** Steps to reproduce the behaviour: 1. Go to 'edit post or page' 2. Click on '3-dotted menu > code editor` 3. Paste the mark up below 4. Click outside the editor (focusing out) 5. The markup changes automatically 6. Publish it 7. The mark up on front end changes even further ```html <ul> <li> <a href='#'> <h3>Title</h3> </a> </li> </ul> ``` **Expected behaviour** Since the markup added is valid HTML I would expect it to be kept as is. But instead, on focusing out the editor, it is changed to (adjusted indentation for readability) ```html <ul> <li> <a href="#"> <h3>Title</h3> </a> <a href="#"></a> </li> </ul> ``` However, when publishing, it changes the markup even further to: ```html <ul> <li> <a href="#"> </p> <h3>Title</h3> <p> </a> <a href="#"></a> </li> </ul> ``` **Additional context** - WordPress 5.3
non_test
code editor not accepting block element inside anchor tag describe the bug whenever editing a page post through the code editor instead of the visual one and adding a block element inside an anchor tag whenever focusing out or saving the page the editor automatically changes the code creating an orphan anchor tag to reproduce steps to reproduce the behaviour go to edit post or page click on dotted menu code editor paste the mark up below click outside the editor focusing out the markup changes automatically publish it the mark up on front end changes even further html title expected behaviour since the markup added is valid html i would expect it to be kept as is but instead on focusing out the editor it is changed to adjusted indentation for readability html title however when publishing it changes the markup even further to html title additional context wordpress
0
216,958
16,825,618,970
IssuesEvent
2021-06-17 18:08:34
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Expose the output of the EE resolver in the job template detail API endpoints
component:api priority:medium state:needs_test
##### ISSUE TYPE - Feature Idea ##### SUMMARY It seems like it would be useful and would mitigate some confusion to expose which EE would be used when executing a template that doesn't have an EE explicitly assigned. **Clarification**: template detail endpoints.
1.0
Expose the output of the EE resolver in the job template detail API endpoints - ##### ISSUE TYPE - Feature Idea ##### SUMMARY It seems like it would be useful and would mitigate some confusion to expose which EE would be used when executing a template that doesn't have an EE explicitly assigned. **Clarification**: template detail endpoints.
test
expose the output of the ee resolver in the job template detail api endpoints issue type feature idea summary it seems like it would be useful and would mitigate some confusion to expose which ee would be used when executing a template that doesn t have an ee explicitly assigned clarification template detail endpoints
1
462,674
13,251,419,707
IssuesEvent
2020-08-20 02:06:33
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
closed
Get stats on time spent on tasks during Case Decision Review
Priority: Low Product: caseflow-queue Stakeholder: BVA Team: Echo 🐬 Type: Metrics or Reporting
## Description Get statistics on how much time users' spend on [tasks associated with Case Decision Review](https://github.com/department-of-veterans-affairs/appeals-team/blob/master/Project%20Folders/Tasks/tasktrees/descr/tasks-overview.md#decision-phase). Getting actual time spent working on task will be challenging, so get as close an estimate as possible using `AppealViews`, `Task.assigned_at`, `updated_at`, `closed_at` and fields in other associated models (e.g., timestamps in `RequestIssue`, `DecisionIssue`, `AttorneyCaseReview`, `JudgeCaseReview`, but ignore `DocumentViews` and other Document-related models as this will be addressed in a different ticket). See #12455. For other ideas, see [Caseflow Metrics](https://docs.google.com/spreadsheets/d/1Q3GYUu5WGwOZErZqGIwwS4OtWIxqOWKTWxBQVcnQ9mo/edit#gid=0). Consider using Google Analytics or other user activity logs. ## Acceptance criteria - [ ] Report containing stats on time spent on tasks associated with Case Decision Review - [ ] If appropriate, identify more accurate measurements or logging that can be implemented ## Background/context/resources To understand where we can improve timeliness of appeals processing, we need to track how users are spending their time in Caseflow. This ticket is a first dip into measuring this and identifying shortcomings, scoped only to the Case Decision Review process. ## Technical notes
1.0
Get stats on time spent on tasks during Case Decision Review - ## Description Get statistics on how much time users' spend on [tasks associated with Case Decision Review](https://github.com/department-of-veterans-affairs/appeals-team/blob/master/Project%20Folders/Tasks/tasktrees/descr/tasks-overview.md#decision-phase). Getting actual time spent working on task will be challenging, so get as close an estimate as possible using `AppealViews`, `Task.assigned_at`, `updated_at`, `closed_at` and fields in other associated models (e.g., timestamps in `RequestIssue`, `DecisionIssue`, `AttorneyCaseReview`, `JudgeCaseReview`, but ignore `DocumentViews` and other Document-related models as this will be addressed in a different ticket). See #12455. For other ideas, see [Caseflow Metrics](https://docs.google.com/spreadsheets/d/1Q3GYUu5WGwOZErZqGIwwS4OtWIxqOWKTWxBQVcnQ9mo/edit#gid=0). Consider using Google Analytics or other user activity logs. ## Acceptance criteria - [ ] Report containing stats on time spent on tasks associated with Case Decision Review - [ ] If appropriate, identify more accurate measurements or logging that can be implemented ## Background/context/resources To understand where we can improve timeliness of appeals processing, we need to track how users are spending their time in Caseflow. This ticket is a first dip into measuring this and identifying shortcomings, scoped only to the Case Decision Review process. ## Technical notes
non_test
get stats on time spent on tasks during case decision review description get statistics on how much time users spend on getting actual time spent working on task will be challenging so get as close an estimate as possible using appealviews task assigned at updated at closed at and fields in other associated models e g timestamps in requestissue decisionissue attorneycasereview judgecasereview but ignore documentviews and other document related models as this will be addressed in a different ticket see for other ideas see consider using google analytics or other user activity logs acceptance criteria report containing stats on time spent on tasks associated with case decision review if appropriate identify more accurate measurements or logging that can be implemented background context resources to understand where we can improve timeliness of appeals processing we need to track how users are spending their time in caseflow this ticket is a first dip into measuring this and identifying shortcomings scoped only to the case decision review process technical notes
0
230,723
18,685,462,744
IssuesEvent
2021-11-01 11:52:15
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
研发商店插件的YAML展示 兼容插件配置错误展示
kind/bug stage/uat stage/test area/ci/frontend area/ci/backend test/passed uat/passed
用户在配置task.json 时,若出现如图 ![image](https://user-images.githubusercontent.com/88175075/138628979-2e6e5d12-ce00-405e-a3da-f621c7795c5b.png) 情况配置成了 ![image](https://user-images.githubusercontent.com/88175075/138629015-4fb9ca9c-4f71-4273-b5e6-d3da7a1f70f0.png) 则会出现插件市场页面显示错误如图 ![image](https://user-images.githubusercontent.com/88175075/138629060-e12cf9ae-939d-464c-a8c5-357818b7f62e.png) 目前期望在 ![image](https://user-images.githubusercontent.com/88175075/138629109-58f96241-d83b-459a-8beb-6012e892d054.png) 如图指定区域提醒用户task.json 配置格式有误,并且用户进入查看页面时,页面不要弹框报错。
2.0
研发商店插件的YAML展示 兼容插件配置错误展示 - 用户在配置task.json 时,若出现如图 ![image](https://user-images.githubusercontent.com/88175075/138628979-2e6e5d12-ce00-405e-a3da-f621c7795c5b.png) 情况配置成了 ![image](https://user-images.githubusercontent.com/88175075/138629015-4fb9ca9c-4f71-4273-b5e6-d3da7a1f70f0.png) 则会出现插件市场页面显示错误如图 ![image](https://user-images.githubusercontent.com/88175075/138629060-e12cf9ae-939d-464c-a8c5-357818b7f62e.png) 目前期望在 ![image](https://user-images.githubusercontent.com/88175075/138629109-58f96241-d83b-459a-8beb-6012e892d054.png) 如图指定区域提醒用户task.json 配置格式有误,并且用户进入查看页面时,页面不要弹框报错。
test
研发商店插件的yaml展示 兼容插件配置错误展示 用户在配置task json 时,若出现如图 情况配置成了 则会出现插件市场页面显示错误如图 目前期望在 如图指定区域提醒用户task json 配置格式有误,并且用户进入查看页面时,页面不要弹框报错。
1
165,901
12,883,125,492
IssuesEvent
2020-07-12 20:22:10
PyTorchLightning/pytorch-lightning
https://api.github.com/repos/PyTorchLightning/pytorch-lightning
closed
Enable Deepsource checks in CI/CD
enhancement help wanted tests / CI wontfix
In addition to flake8, mypy and other checks, I use https://deepsource.io/ for my projects. It is free and from time to time it finds non-trivial bugs. When I use it for projects with Pytorch-lightning and several validation loaders it throws me an error in python style and syntax. To overcome this I need to add things like: `# skipcq: PYL-W0201` to my code. I would prefer to avoid it. Is it possible that Deepsource checks would be allowed in Pytorch Lightning repo? I believe it will lead to better code quality.
1.0
Enable Deepsource checks in CI/CD - In addition to flake8, mypy and other checks, I use https://deepsource.io/ for my projects. It is free and from time to time it finds non-trivial bugs. When I use it for projects with Pytorch-lightning and several validation loaders it throws me an error in python style and syntax. To overcome this I need to add things like: `# skipcq: PYL-W0201` to my code. I would prefer to avoid it. Is it possible that Deepsource checks would be allowed in Pytorch Lightning repo? I believe it will lead to better code quality.
test
enable deepsource checks in ci cd in addition to mypy and other checks i use for my projects it is free and from time to time it finds non trivial bugs when i use it for projects with pytorch lightning and several validation loaders it throws me an error in python style and syntax to overcome this i need to add things like skipcq pyl to my code i would prefer to avoid it is it possible that deepsource checks would be allowed in pytorch lightning repo i believe it will lead to better code quality
1
206,945
15,784,749,271
IssuesEvent
2021-04-01 15:29:52
qauser21/backup3
https://api.github.com/repos/qauser21/backup3
opened
ABAC_Level3 on PUT:/api/v1/primary-transaction
CVSS_3.1 8.1 GitBack Test 3 High apisec
Title: ABAC_Level3 Vulnerability on PUT:/api/v1/primary-transaction Project: GitBack Test 3 Description: The ABAC exploit allows an attacker to read, modify, delete, add and perform actions on customer/un-authorized data. <style>strong { color: #ef5350!important;}</style><strong>Assertion </strong> <b><font color="#ef5350">Name: </font></b> Attribute Based Access Control 3 (ABAC 3) <b><font color="#ef5350"><sup>( 1 )</sup></font></b><br> <br> <b><font color="#ef5350">Overview: </font></b> Access Control (or Authorization) is the process of granting or denying specific requests from a user, program, or process. Access control also involves the act of granting and revoking those privileges. Attribute Based Access Control (ABAC) will grant or deny user requests based on arbitrary attributes of the user and arbitrary attributes of the object, and environment conditions that may be globally recognized and more relevant to the policies at hand. <br> <br>'Attribute-based-access-control 3' identifies dependent and nested resource/data/record vulnerabilities. e.g. vulnerabilities in tasks which is nested within a project and the access-controls may only be applied at the project level. ABAC scanning identifies data/resource leak/attack vulnerabilities. Looks for private user/account data being illegally read, written, updated, deleted or operated by other users or tenants or accounts.<br> <br>This scanner requires a private-account/user to create private data/resources e.g. UserA. And it also requires other users who shouldn't have any access to UserA's data like UserB, UserC, & UserD based on your App multi-tenancy model. e.g. UserA can be a user from tenant/org-a and UserB can be a user in tenant-b and UserC can be a user in tenant-c with admin privileges.<br /> <br />APIs tend to expose endpoints that handle object identifiers, creating a wide attack surface Level Access Control issue. Object-level authorization checks should be considered in every function that accesses a data source using input from the user.<br> <br> <b><font color="#ef5350">Severity: </font></b> OWASP 2019 API Top 10 ranks ABAC vulnerabilities at Top 1 position and is named Broken Object Level Authorization. <b><font color="#ef5350"><sup>( 2 )</sup></font></b> <br> <br> <b><font color="#ef5350">Vulnerability Impact: </font></b> With flawed or broken ABAC security control policy in place, The following are some of the consequences. <br> <br> <ul> <li>Unauthorized access can result in data disclosure to unauthorized parties, data loss, or data manipulation.</li> <li>Unauthorized access to objects can also lead to full account takeover.</li> </ul> <b><font color="#ef5350">Exploitation: </font></b> Attackers can exploit API endpoints that are vulnerable to broken object level authorization by manipulating the ID of an object that is sent within the request. This may lead to unauthorized access to sensitive data. This issue is extremely common in API-based applications because the server component usually does not fully track the client’s state, and instead, relies more on parameters like object IDs, that are sent from the client to decide which objects to access. <br> <br> <b><font color="#ef5350">Remediation: </font></b> The following techniques may be checked for ensuring RBAC is in place <b><font color="#ef5350"><sup>( 2 ) ( 3 ) ( 4 )</sup></font></b>. <br> <br> <ul> <li>Implement a proper authorization mechanism that relies on the user policies and hierarchy.</li> <li>Prefer not to use an ID that has been sent from the client, but instead use an ID that is stored in the session object when accessing a database record by the record ID.</li> <li>Use an authorization mechanism to check if the logged-in user has access to perform the requested action on the record in every function that uses an client input to access a record in the database.</li> <li>Prefer to use random and unpredictable values as GUIDs for records’ IDs.</li> <li>Write tests to evaluate the authorization mechanism. Do not deploy vulnerable changes that break the tests.</li> </ul> <b><font color="#ef5350">References: </font></b><br> <ol> <li>Enforce Access Controls - https://owasp-top-10-proactive-controls-2018.readthedocs.io/en/latest/c7-enforce-access-controls.html</li> <li>OWASP 2019 API Top 10 - https://github.com/OWASP/API-Security/raw/master/2019/en/dist/owasp-api-security-top-10.pdf</li> <li>OWASP Access Control Cheat Sheet - https://cheatsheetseries.owasp.org/cheatsheets/Access_Control_Cheat_Sheet.html</li> <li>OWASP REST Security Cheat Sheet - https://cheatsheetseries.owasp.org/cheatsheets/REST_Security_Cheat_Sheet.html </li> </ol> <br>Risk: ABAC_Level3 Severity: High API Endpoint: http://95.217.118.53:8080/api/v1/primary-transaction Environment: Master Playbook: ApiV1PrimaryTransactionPutPrimarytransactionuserbDisallowAbact3 Researcher: [apisec Bot] QUICK TIPS Suggestion: Add access-control checks on incoming requests against all data calls. Effort Estimate: 2.0 Wire Logs: IMPORTANT LINKS Vulnerability Details: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/recommendations/8a8083b8788dc46901788e0df69f32d4/details Project: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/jobs Environment: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/environments/8a8083b8788dc46901788e09055921c9/edit Scan Dashboard: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/jobs/8a8083b8788dc46901788e090fe32356/runs/8a8083b8788dc46901788e0dcfa532aa Playbook: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/template/ApiV1PrimaryTransactionPutPrimarytransactionuserbDisallowAbact3 Coverage: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/configuration Code Sample: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/recommendations/8a8083b8788dc46901788e0df69f32d4/codesamples PS: Please contact support@apisec.ai for apisec access and login issues. --- apisec Bot ---
1.0
ABAC_Level3 on PUT:/api/v1/primary-transaction - Title: ABAC_Level3 Vulnerability on PUT:/api/v1/primary-transaction Project: GitBack Test 3 Description: The ABAC exploit allows an attacker to read, modify, delete, add and perform actions on customer/un-authorized data. <style>strong { color: #ef5350!important;}</style><strong>Assertion </strong> <b><font color="#ef5350">Name: </font></b> Attribute Based Access Control 3 (ABAC 3) <b><font color="#ef5350"><sup>( 1 )</sup></font></b><br> <br> <b><font color="#ef5350">Overview: </font></b> Access Control (or Authorization) is the process of granting or denying specific requests from a user, program, or process. Access control also involves the act of granting and revoking those privileges. Attribute Based Access Control (ABAC) will grant or deny user requests based on arbitrary attributes of the user and arbitrary attributes of the object, and environment conditions that may be globally recognized and more relevant to the policies at hand. <br> <br>'Attribute-based-access-control 3' identifies dependent and nested resource/data/record vulnerabilities. e.g. vulnerabilities in tasks which is nested within a project and the access-controls may only be applied at the project level. ABAC scanning identifies data/resource leak/attack vulnerabilities. Looks for private user/account data being illegally read, written, updated, deleted or operated by other users or tenants or accounts.<br> <br>This scanner requires a private-account/user to create private data/resources e.g. UserA. And it also requires other users who shouldn't have any access to UserA's data like UserB, UserC, & UserD based on your App multi-tenancy model. e.g. UserA can be a user from tenant/org-a and UserB can be a user in tenant-b and UserC can be a user in tenant-c with admin privileges.<br /> <br />APIs tend to expose endpoints that handle object identifiers, creating a wide attack surface Level Access Control issue. Object-level authorization checks should be considered in every function that accesses a data source using input from the user.<br> <br> <b><font color="#ef5350">Severity: </font></b> OWASP 2019 API Top 10 ranks ABAC vulnerabilities at Top 1 position and is named Broken Object Level Authorization. <b><font color="#ef5350"><sup>( 2 )</sup></font></b> <br> <br> <b><font color="#ef5350">Vulnerability Impact: </font></b> With flawed or broken ABAC security control policy in place, The following are some of the consequences. <br> <br> <ul> <li>Unauthorized access can result in data disclosure to unauthorized parties, data loss, or data manipulation.</li> <li>Unauthorized access to objects can also lead to full account takeover.</li> </ul> <b><font color="#ef5350">Exploitation: </font></b> Attackers can exploit API endpoints that are vulnerable to broken object level authorization by manipulating the ID of an object that is sent within the request. This may lead to unauthorized access to sensitive data. This issue is extremely common in API-based applications because the server component usually does not fully track the client’s state, and instead, relies more on parameters like object IDs, that are sent from the client to decide which objects to access. <br> <br> <b><font color="#ef5350">Remediation: </font></b> The following techniques may be checked for ensuring RBAC is in place <b><font color="#ef5350"><sup>( 2 ) ( 3 ) ( 4 )</sup></font></b>. <br> <br> <ul> <li>Implement a proper authorization mechanism that relies on the user policies and hierarchy.</li> <li>Prefer not to use an ID that has been sent from the client, but instead use an ID that is stored in the session object when accessing a database record by the record ID.</li> <li>Use an authorization mechanism to check if the logged-in user has access to perform the requested action on the record in every function that uses an client input to access a record in the database.</li> <li>Prefer to use random and unpredictable values as GUIDs for records’ IDs.</li> <li>Write tests to evaluate the authorization mechanism. Do not deploy vulnerable changes that break the tests.</li> </ul> <b><font color="#ef5350">References: </font></b><br> <ol> <li>Enforce Access Controls - https://owasp-top-10-proactive-controls-2018.readthedocs.io/en/latest/c7-enforce-access-controls.html</li> <li>OWASP 2019 API Top 10 - https://github.com/OWASP/API-Security/raw/master/2019/en/dist/owasp-api-security-top-10.pdf</li> <li>OWASP Access Control Cheat Sheet - https://cheatsheetseries.owasp.org/cheatsheets/Access_Control_Cheat_Sheet.html</li> <li>OWASP REST Security Cheat Sheet - https://cheatsheetseries.owasp.org/cheatsheets/REST_Security_Cheat_Sheet.html </li> </ol> <br>Risk: ABAC_Level3 Severity: High API Endpoint: http://95.217.118.53:8080/api/v1/primary-transaction Environment: Master Playbook: ApiV1PrimaryTransactionPutPrimarytransactionuserbDisallowAbact3 Researcher: [apisec Bot] QUICK TIPS Suggestion: Add access-control checks on incoming requests against all data calls. Effort Estimate: 2.0 Wire Logs: IMPORTANT LINKS Vulnerability Details: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/recommendations/8a8083b8788dc46901788e0df69f32d4/details Project: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/jobs Environment: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/environments/8a8083b8788dc46901788e09055921c9/edit Scan Dashboard: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/jobs/8a8083b8788dc46901788e090fe32356/runs/8a8083b8788dc46901788e0dcfa532aa Playbook: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/template/ApiV1PrimaryTransactionPutPrimarytransactionuserbDisallowAbact3 Coverage: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/configuration Code Sample: https://95.217.198.167/#/app/projects/8a8083b8788dc46901788e09054221c6/recommendations/8a8083b8788dc46901788e0df69f32d4/codesamples PS: Please contact support@apisec.ai for apisec access and login issues. --- apisec Bot ---
test
abac on put api primary transaction title abac vulnerability on put api primary transaction project gitback test description the abac exploit allows an attacker to read modify delete add and perform actions on customer un authorized data strong color important assertion name attribute based access control abac overview access control or authorization is the process of granting or denying specific requests from a user program or process access control also involves the act of granting and revoking those privileges attribute based access control abac will grant or deny user requests based on arbitrary attributes of the user and arbitrary attributes of the object and environment conditions that may be globally recognized and more relevant to the policies at hand attribute based access control identifies dependent and nested resource data record vulnerabilities e g vulnerabilities in tasks which is nested within a project and the access controls may only be applied at the project level abac scanning identifies data resource leak attack vulnerabilities looks for private user account data being illegally read written updated deleted or operated by other users or tenants or accounts this scanner requires a private account user to create private data resources e g usera and it also requires other users who shouldn t have any access to usera s data like userb userc userd based on your app multi tenancy model e g usera can be a user from tenant org a and userb can be a user in tenant b and userc can be a user in tenant c with admin privileges apis tend to expose endpoints that handle object identifiers creating a wide attack surface level access control issue object level authorization checks should be considered in every function that accesses a data source using input from the user severity owasp api top ranks abac vulnerabilities at top position and is named broken object level authorization vulnerability impact with flawed or broken abac security control policy in place the following are some of the consequences unauthorized access can result in data disclosure to unauthorized parties data loss or data manipulation unauthorized access to objects can also lead to full account takeover exploitation attackers can exploit api endpoints that are vulnerable to broken object level authorization by manipulating the id of an object that is sent within the request this may lead to unauthorized access to sensitive data this issue is extremely common in api based applications because the server component usually does not fully track the client’s state and instead relies more on parameters like object ids that are sent from the client to decide which objects to access remediation the following techniques may be checked for ensuring rbac is in place implement a proper authorization mechanism that relies on the user policies and hierarchy prefer not to use an id that has been sent from the client but instead use an id that is stored in the session object when accessing a database record by the record id use an authorization mechanism to check if the logged in user has access to perform the requested action on the record in every function that uses an client input to access a record in the database prefer to use random and unpredictable values as guids for records’ ids write tests to evaluate the authorization mechanism do not deploy vulnerable changes that break the tests references enforce access controls owasp api top owasp access control cheat sheet owasp rest security cheat sheet risk abac severity high api endpoint environment master playbook researcher quick tips suggestion add access control checks on incoming requests against all data calls effort estimate wire logs important links vulnerability details project environment scan dashboard playbook coverage code sample ps please contact support apisec ai for apisec access and login issues apisec bot
1
271,407
23,602,627,581
IssuesEvent
2022-08-24 04:36:53
woocommerce/woocommerce-ios
https://api.github.com/repos/woocommerce/woocommerce-ios
closed
[UI Testing] - Re-enable skipped login test cases
type: task feature: login category: ui tests
When the gear icon was moved from My Store to the Settings screen, some of the login-related test cases were skipped because of the screen change: https://github.com/woocommerce/woocommerce-ios/blob/65eed594fb215341616722158b4973335ad6a69f/WooCommerce/WooCommerceUITests/Tests/LoginTests.swift#L15-L54 This is to track the work to re-enable the tests.
1.0
[UI Testing] - Re-enable skipped login test cases - When the gear icon was moved from My Store to the Settings screen, some of the login-related test cases were skipped because of the screen change: https://github.com/woocommerce/woocommerce-ios/blob/65eed594fb215341616722158b4973335ad6a69f/WooCommerce/WooCommerceUITests/Tests/LoginTests.swift#L15-L54 This is to track the work to re-enable the tests.
test
re enable skipped login test cases when the gear icon was moved from my store to the settings screen some of the login related test cases were skipped because of the screen change this is to track the work to re enable the tests
1
114,792
24,664,748,615
IssuesEvent
2022-10-18 09:26:40
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
JIT: GDV reorders null checks of 'this' with arguments
area-CodeGen-coreclr in-pr
The following example throws NRE without printing any output when tiered PGO is enabled. It should print "Should be printed" before throwing NRE. ```csharp using System; using System.Runtime.CompilerServices; using System.Threading; public class Program { public static void Main() { for (int i = 0; i < 100; i++) { long sum = Foo(i => i * 42, true); if (i > 30 && i < 40) Thread.Sleep(100); } Foo(null, false); } [MethodImpl(MethodImplOptions.NoInlining)] private static long Foo(Func<long, long> f, bool silent) { long result = 0; for (long i = 0; i < 100000; i++) result += f(Test(i, silent)); return result; } [MethodImpl(MethodImplOptions.NoInlining)] private static long Test(long i, bool silent) { if (i == 0 && !silent) Console.WriteLine("Should be printed"); return i; } } ``` The evaluation order is meant to be: 1. Evaluate this 2. Evaluate arguments 3. Null check this 4. Do call With GDV, since the guard involves dereferencing 'this', step 3 happens too early. This happens for both vtable, interface and delegate GDV. Tricky to solve unfortunately. We need to evaluate the guard after the arguments (hard to represent without lots of additional spilling) or add a null check to the guard (expensive). cc @AndyAyersMS @EgorBo
1.0
JIT: GDV reorders null checks of 'this' with arguments - The following example throws NRE without printing any output when tiered PGO is enabled. It should print "Should be printed" before throwing NRE. ```csharp using System; using System.Runtime.CompilerServices; using System.Threading; public class Program { public static void Main() { for (int i = 0; i < 100; i++) { long sum = Foo(i => i * 42, true); if (i > 30 && i < 40) Thread.Sleep(100); } Foo(null, false); } [MethodImpl(MethodImplOptions.NoInlining)] private static long Foo(Func<long, long> f, bool silent) { long result = 0; for (long i = 0; i < 100000; i++) result += f(Test(i, silent)); return result; } [MethodImpl(MethodImplOptions.NoInlining)] private static long Test(long i, bool silent) { if (i == 0 && !silent) Console.WriteLine("Should be printed"); return i; } } ``` The evaluation order is meant to be: 1. Evaluate this 2. Evaluate arguments 3. Null check this 4. Do call With GDV, since the guard involves dereferencing 'this', step 3 happens too early. This happens for both vtable, interface and delegate GDV. Tricky to solve unfortunately. We need to evaluate the guard after the arguments (hard to represent without lots of additional spilling) or add a null check to the guard (expensive). cc @AndyAyersMS @EgorBo
non_test
jit gdv reorders null checks of this with arguments the following example throws nre without printing any output when tiered pgo is enabled it should print should be printed before throwing nre csharp using system using system runtime compilerservices using system threading public class program public static void main for int i i i long sum foo i i true if i i thread sleep foo null false private static long foo func f bool silent long result for long i i i result f test i silent return result private static long test long i bool silent if i silent console writeline should be printed return i the evaluation order is meant to be evaluate this evaluate arguments null check this do call with gdv since the guard involves dereferencing this step happens too early this happens for both vtable interface and delegate gdv tricky to solve unfortunately we need to evaluate the guard after the arguments hard to represent without lots of additional spilling or add a null check to the guard expensive cc andyayersms egorbo
0
53,331
11,037,677,655
IssuesEvent
2019-12-08 08:17:05
pywbem/pywbem
https://api.github.com/repos/pywbem/pywbem
closed
MOF compiler: Does not raise errors from ModifyClass when compiling an existing class
area: code resolution: fixed roll back/forward done type: bug
When the MOF compiler compiles a class, it invokes `p_mp_createClass()` which invokes `handle.CreateClass()` and falls back to `handle.ModifyClass()` if the `handle.CreateClass()` method returns that the class already exists. The issue is that if the `handle.ModifyClass()` call raises a CIMError, that error is logged but not re-raised. I verified the code back to branch stable_0.8 and it never raised such an error. The situation is complicated by the fact that `MOFWBEMConnection.CreateClass()` already replaces the class if it already exists, so that method never returns that the class already exists. Issue #991 describes that behavior. So if the MOFCompiler handle is a `MOFWBEMConnection` object, the issue described above does not have any bad consequences. However, the MOFCompiler handle does not have to be a `MOFWBEMConnection` object, so in the general case (e.g. with a real WBEM server), the issue does have bad consequences. Proposal: * Change `MOFWBEMConnection.CreateClass()` to no longer replace an existing class, but to return that it exists. This fixes issue #991. * Change `p_mp_createClass()` to raise an error if `handle.ModifyClass()` returns an error. * Add according testcases.
1.0
MOF compiler: Does not raise errors from ModifyClass when compiling an existing class - When the MOF compiler compiles a class, it invokes `p_mp_createClass()` which invokes `handle.CreateClass()` and falls back to `handle.ModifyClass()` if the `handle.CreateClass()` method returns that the class already exists. The issue is that if the `handle.ModifyClass()` call raises a CIMError, that error is logged but not re-raised. I verified the code back to branch stable_0.8 and it never raised such an error. The situation is complicated by the fact that `MOFWBEMConnection.CreateClass()` already replaces the class if it already exists, so that method never returns that the class already exists. Issue #991 describes that behavior. So if the MOFCompiler handle is a `MOFWBEMConnection` object, the issue described above does not have any bad consequences. However, the MOFCompiler handle does not have to be a `MOFWBEMConnection` object, so in the general case (e.g. with a real WBEM server), the issue does have bad consequences. Proposal: * Change `MOFWBEMConnection.CreateClass()` to no longer replace an existing class, but to return that it exists. This fixes issue #991. * Change `p_mp_createClass()` to raise an error if `handle.ModifyClass()` returns an error. * Add according testcases.
non_test
mof compiler does not raise errors from modifyclass when compiling an existing class when the mof compiler compiles a class it invokes p mp createclass which invokes handle createclass and falls back to handle modifyclass if the handle createclass method returns that the class already exists the issue is that if the handle modifyclass call raises a cimerror that error is logged but not re raised i verified the code back to branch stable and it never raised such an error the situation is complicated by the fact that mofwbemconnection createclass already replaces the class if it already exists so that method never returns that the class already exists issue describes that behavior so if the mofcompiler handle is a mofwbemconnection object the issue described above does not have any bad consequences however the mofcompiler handle does not have to be a mofwbemconnection object so in the general case e g with a real wbem server the issue does have bad consequences proposal change mofwbemconnection createclass to no longer replace an existing class but to return that it exists this fixes issue change p mp createclass to raise an error if handle modifyclass returns an error add according testcases
0
57,472
14,163,296,086
IssuesEvent
2020-11-12 02:00:59
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
GetExternalLoginInfoAsync is null when called from ExternalLoginModel.OnGetCallbackAsync - IFrames involved.
area-security
Goal: Perform a silent login using a hidden iframe to google. Result: I can achieve this by passing prompt=none to google which then doesn't x-frame-options=deny me. All Good when I run the app normally and I route to ``` https://localhost:5001/Plugin ``` and ``` https://127.0.0.1.xip.io:5001/Plugin ``` My Plugin Page has my hidden iFrame. ``` <button id='signin'>signin</button> <p id="received-message">I've heard nothing yet</p> <form style="display: none;" id="loginform" target="guestFrame" action="/Identity/Account/ExternalLogin?returnUrl=%2FOIDCIFrameResult&amp;errorUrl=%2FOIDCIFrameResult&amp;prompt=none&amp;provider=google" method="post"> <div> <p> <input type="submit"> </p> </div> </form> <iframe style="display: none;" id="guestFrame" name="guestFrame" src="/Blank"> Your browser does not support inline frames. </iframe> ``` Ok, now put the entire app in an iFrame and try to do a silent login under that context. ``` <iframe id="guestFrame" name="guestFrame" src="https://127.0.0.1.xip.io:5001/Plugin"> </iframe> ``` The above ```iFrame``` is hosted in ```https://localhost:5001/IFrameHost``` I am using [xip.io](https://xip.io) as my wildcard DNS. ```https://127.0.0.1.xip.io:5001``` My ```signin-google``` gets hit and it looks like there is a good code in there. ``` GENERAL Request URL: https://127.0.0.1.xip.io:5001/signin-google Request Method: POST Status Code: 302 Remote Address: 127.0.0.1:5001 Referrer Policy: no-referrer-when-downgrade Response Headers cache-control: no-cache content-length: 0 date: Fri, 28 Aug 2020 20:51:15 GMT expires: Thu, 01 Jan 1970 00:00:00 GMT location: /Identity/Account/ExternalLogin?returnUrl=%2FOIDCIFrameResult&handler=Callback pragma: no-cache server: Kestrel set-cookie: .AspNetCore.Correlation.google.gxGuRZ5ZVbOr4SmC3zd6CLQpHZ_j-ZTc-jyvSnAF-IQ=; expires=Thu, 01 Jan 1970 00:00:00 GMT; path=/signin-google; secure; samesite=none; httponly set-cookie: .AspNetCore.OpenIdConnect.Nonce.CfDJ8L16rcoVAjJMt5TusKivrFcSEujxcxPIbAwED3uPgd2mk-J_fEJKkogtr3NaSVolWh5Rv5DcehuReLphIyvY7VG-iypvHqGc7knVZXv2FWJaFed1bQaA2jkGFbmEmQs4DydnmwOdlA4Df_o8SjCPEW8Bq54SWJz6uNI8L9p0_e8ZllW1gaF0zCOuoQn_XdEm8jUW6o1h2N1o9YLFAbj3-w0rj5BWqOfl6ysEkeY0DXe7xm8Ycmd7PpC98edYtnkUgs5Kfs7H_4bn0E4TSP_1XS8=; expires=Thu, 01 Jan 1970 00:00:00 GMT; path=/signin-google; secure; samesite=none; httponly set-cookie: Identity.External=CfDJ8L16rcoVAjJMt5TusKivrFecjgIwf31r3SrfRXOwnrnUe9wLJ38mJtd8HAzKOc_2rsgrFsLc6lAnKcHa5pYmafjUFaKJcmBZt2ys2JSUl3qFKVqSTZaSpS4E21Tt70yaz2T9AzSzDOhBv2hIBrtj1Y9c1Ximra9zPlTXLI1PuUNPl7ZQSTzQfep7-hhr7Fj0aCz4Qq14UvjHBaU8QjGd_XACdODJc4aX1PlT8_JgXDl5ob9RtDxmta1egpT6EkFOFwBC1Ngb0UiY5Ilz-c1R3255SDIxa6PVVNjy55ccFo0Hpfik9JWj6oTgGApt6koNrxbFOTK5sLAt5S7lR1FbL8Jkq_DOpmTzfSgEPVV7PIfW6AK4zZav6qawqgTn01mYRM-NO8f_gcrhTBfQpEbrYAFqVTu-Svo6ZkKTI-iRMe57arPz8e8nUvRnqplAvR8MbRKvXNAHGdRcN-JDvzxZQHxm7I0u-gX0MUm1CezYDhbaENhosyZrcxZhNZ9u7XdVgEtAlPe_KX1Vdn0zHmgHadcW_q8l1yvVzMTigryWr8-6wvGyeCPVhGGfRH6vIqkF9hGgTNXLpIso3xESmRJveCXbYWZfZ9Tm2w51T4ToCwOc5ymNmYkUBvQYX0HCpKkwOIlCn1JLGWy7q6WcN1VRcZHImvX-TA3R9aiIqRwbuAe3Ju4Qop8oAGoXW1FxFPAGIQRytj_m7uPBdngitb5bZgcFKhXr5vgxngWrE2Ns55ItXu85aGHi8DymfUmuH-Wtsd6haLpq39E08D_zudyY11hbt2_raO9OPNkWV8R_QCYR4eij7sp1OTG_Jq1tkdsQRTe0DfmxE9-reYYw8hK_sF4Hb8GU8QO0Zrn0Ia9UrF6tqW-LNhO5Id02Jmxl2Yn6yOipNwO9FbRDmLy9O4LumZWZw1miLS-XJ9KxPRPcR3x_wCONtKrxfWZzEqYs7T34CSQB71okiKFwfMu_vUNJl44TUsmgZ9QEtgIxlHcSeIHPc7Y2FYU3ZEM3gBmekbGGywzQ48TxNedsHo9PJ_We5BHbyBQK5XvVJ8r22VBOYlZHaNbcOt3WS7KxJp-uHmlfCY7NuC5d27IqVN0Ir5sLDbnr6KSblI2y82faQXoXUQz-HQlTly4O6lEJc1YbEAcfgHWieye-3A9wbjEt6N7HISTjcWWNgJIk_IfJcQC2lP95wP5VHZgwKQrIitkRCYaUxwfTEqUzVZq34chEd9Y8eF1A-LAJjHpT4nsovrRuD1qof63XmJ-dqwdAEeC41CLRDgWRZRdAsFj2xEZkTIDu8IQZugOfxucgYBH8TxWIssU7YLtR4CVQoDOCmT-EIodwEPAY0Bi_mMHh6O0EhrGyXsvEFFm_M-A-6uLZgrL8jjU2DNUG9ZhFSIUncDZ1_9iEvuZfv8OMUCWayEmmml8-SXeElPvZTDk4O2Els11JRP7q1jTHZ-5vakdbRDql68HiHryY8xEsDN_4LW_blaHCwKV32npzrB_vnETmj0Ii7M1QytZV3Nfdr8R-bfYg6wF46z5bm-9TwWkX4_pwJ5dp116kvE5LkADo-NIrNo33EJCsp-iYYhmUWRVL5Rh7nszVHr-1fj2TramHcKRn7_oi4djcFIHl0w-i4dF3kRPy-dgE2nKHsSM9Rpyjn-whNSQfPM7AOxIYTbllejAmGVJxUU83TIrQiOeNxml6e5mlp3GrVmC5qtBmnUA4hdrpyGgWFzXx5uBCtLN2z39HrrKvIKD25naRPqu6mRHI5QRqSFL9mpmv4Cseqk975MnuUPtjI-8BhpGoavN_4PALb_VeM-JTh3TW0W40S6BOi3c97ZojY_J34EEO-WSk5zNw1WHS233etkQ5qCsjtae2f3_dC9Gi_SI9CgFgEnz6LjqqaXOisgVGJTGC-D4V-Ms0teHoEklFYrvhUMZiCmxGdW8FwrZF6bG9uAKX1iRvg_557SIczHx24v9zXZL33dN0m11THnnD_0yVAc6ac9uiXBWmI0eZ-0jxk0Rg2dZsZ75RS-Mj8CNQZPfPfvfQj9LpZh8uyS3c3IPJVLzLyVpmEC0co-8gty45-qFU-cjoyiPI3v9hH9HWcO6bcdp7w1dgj-nQxPCj0Sb8MkPbW2cvl2y9Hwr5MmAjEp8gnCtLlfvm7hbbzwntkQURVP2V7ir83ZjNfObKcXN7SPbcsf88yTNfnilrLyF-GfaufOiqz5PEEy33k2geQMRbjbjdzpEarQjLZnQGLZ5hupMmbr7UdME2e1PCZPBp-KWZP4qA8uW3gCsmSqhBQgo4fcmjFocPPyt1IEnhmJqfI8c32_CvCK8Df98f183xwuBBVJr23MN1rVYf4avXEnrTfrs6I6CJfy8ZN5upxcUqpMxTQiMcO1lVycx8WYvcZTl1VXc9CipLiE7t92rqLeoyUQENWSvryTIL5JDtEexJTZ5fGDyLc1NCq5JidkG-oH32aPtSwm1hEQa8MsWfK64N_8f5YA5KVQuen-d-IQil_u7LVNS5BfIjEuhydZqRdoEi-8IKoraDCFzs6prdaXzqB3rQJwa2cOnuCjQNT8xz2kvRBCq75rt7I-srPSYNA57W9SwP5dn8GR2AxUI8CcO7zCfuBwO4FbcEgBtxCWaDmhAVnZd91MH0sslMo4hqNI8PQnNNvtGDXuHsM68tETZflcEW1EEH6Ylr2LWCDGlEZEK6PI1QyYI8TteMw0Rv6dxHbJqhYtvmb3Mx9FIf48NCEWNOnHK6DLt4R6gf-DjZp8pe8z1yMQ8sUccE_4R53irREwFrwij8BUHm1ncDzAVBK8uaGoHskM4oeLjKn8xCEW5pOuOBzXryrt9d0-3ymmLXUsI8du2WB0LmsMOJ3IM4f7zi9P-TjWTt4bdJ7YAmF5lp_zU0ltvMexKB9aTj-b7gfoEd2nBL3rd4E_VCqh_WosaI_t27CulyuuWCsGEFChO1vC0ONeyKHcjJSBv7E3kUYoYvo664SGC8OBZe69VCxFSPOhersp8XS74F6VtTzj7B8xvht8IsuOmZ7Ce5dj6ZuacUarAVppJBzeFfE2HrDOKLQGZ9Irq9berzA9zmQlHFDGNbj3GkZ5x8RmAc-pn1w_GAHeIVzULfNZ8Gxc-3AkqwlqkbDVZEuUL10zgtjh9vezwhiAvllcKS86AunCdb6UKegoSKPHmAWHQBLNwVYBjs5lDBOlxpzN41UhHAAlKmBfBS4QDoejTZN8yZGWS7irf-9VmBDiI1oSh6UbajAXLVmKJ6cE2zqXBjrl_YEpqyPVIIENYDuaa2v4f_t-ErgUQb3QN7jOKVF_QhOru3NQyWcZcBPEUlStTnayT1V0LK_2yjApvG9tLGgpZVGdsza9sbMexUQjRu3phFS1maVP3nV3Ee5D2mWWCzOwqjRvgOObzR4XxY8-9hIDI7iho3mNhpoiNubvH4s2ocmfImHmZqRBuLB-pNZ4dSAAnPIBDLpVWLmLLnxp5zIs2phfI7_GrHAz3LUj1ODVT2lyBjQYI_bNgpBHzfKhNfMBnHm4HpC_d7v7tD4X1cD_dOCBu3VQIs77u60sjOemot7g_F__P9GdrJQC60jZyWnW7IOE8V1Kg1ZtRISRBGavq3kj37H1qLHqEE6_R4tIgUvQiWb-9BoegnTKaX6_Cpq4L_OFGiegOr7EalAf2KcDm1AL15FEcHuYlHe5YeW-XGaHQdG7Ovt2owKAO1Kg; path=/; secure; samesite=lax; httponly status: 302 x-correlation-id: d5a7b7fe-c5e9-4ab0-b6ae-0cb3ba891827 FORM-DATA state=CfDJ8L16rcoVAjJMt5TusKivrFf2GEKRNyWN6Am9QMgqgxz17T9ThyCs-KUxjv1eUpwqojsNxikmLmsC5QXNBURb4gj8IzcQ1lYVPx4iP20oEk3bO2ZQgibTGRDlkBOAp6S7fSCUOPi1lwI1a5H9JjmewSY_4lpEftzbPKn9PCIbHuIkEcbh1UIDkptbanBb-ALt40tmtX8_UzgR7ZQJLLzzNp7POLvKSSXR8qugXU8CefSHCJGGNFM4GnNEVkRVW4oIhm2afXjlLJ13D90yaa8XzVPz_gHAiCgrCM4LVaDqTBfjHWZDiDAyDl66H_XGTaYjyNosKGStsZR3pF-23O0BQrmaXriT6hPTxKQnimMytT3DREAG0G6otxk_mBBMbQziepTeIqNWAMePCEOXZ5Ha367R343Kzyill8D_YNt82uTcFmf9xe-pAT7Mn9Q6P-p0Vsvs5WWN8dQyXposYpUThN-Ub6IoK4WtnNI2gYKBMR_S&code=REDACTED&scope=profile+openid+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.profile&authuser=0&prompt=none ``` ```ExternalLoginModel.OnGetCallbackAsync``` gets called but then I get a ```null``` when calling ```GetExternalLoginInfoAsync ``` I recall that some cookies get cleared during this exchange, but why would simply hosting my app in an iFrame result in different behavior?
True
GetExternalLoginInfoAsync is null when called from ExternalLoginModel.OnGetCallbackAsync - IFrames involved. - Goal: Perform a silent login using a hidden iframe to google. Result: I can achieve this by passing prompt=none to google which then doesn't x-frame-options=deny me. All Good when I run the app normally and I route to ``` https://localhost:5001/Plugin ``` and ``` https://127.0.0.1.xip.io:5001/Plugin ``` My Plugin Page has my hidden iFrame. ``` <button id='signin'>signin</button> <p id="received-message">I've heard nothing yet</p> <form style="display: none;" id="loginform" target="guestFrame" action="/Identity/Account/ExternalLogin?returnUrl=%2FOIDCIFrameResult&amp;errorUrl=%2FOIDCIFrameResult&amp;prompt=none&amp;provider=google" method="post"> <div> <p> <input type="submit"> </p> </div> </form> <iframe style="display: none;" id="guestFrame" name="guestFrame" src="/Blank"> Your browser does not support inline frames. </iframe> ``` Ok, now put the entire app in an iFrame and try to do a silent login under that context. ``` <iframe id="guestFrame" name="guestFrame" src="https://127.0.0.1.xip.io:5001/Plugin"> </iframe> ``` The above ```iFrame``` is hosted in ```https://localhost:5001/IFrameHost``` I am using [xip.io](https://xip.io) as my wildcard DNS. ```https://127.0.0.1.xip.io:5001``` My ```signin-google``` gets hit and it looks like there is a good code in there. ``` GENERAL Request URL: https://127.0.0.1.xip.io:5001/signin-google Request Method: POST Status Code: 302 Remote Address: 127.0.0.1:5001 Referrer Policy: no-referrer-when-downgrade Response Headers cache-control: no-cache content-length: 0 date: Fri, 28 Aug 2020 20:51:15 GMT expires: Thu, 01 Jan 1970 00:00:00 GMT location: /Identity/Account/ExternalLogin?returnUrl=%2FOIDCIFrameResult&handler=Callback pragma: no-cache server: Kestrel set-cookie: .AspNetCore.Correlation.google.gxGuRZ5ZVbOr4SmC3zd6CLQpHZ_j-ZTc-jyvSnAF-IQ=; expires=Thu, 01 Jan 1970 00:00:00 GMT; path=/signin-google; secure; samesite=none; httponly set-cookie: .AspNetCore.OpenIdConnect.Nonce.CfDJ8L16rcoVAjJMt5TusKivrFcSEujxcxPIbAwED3uPgd2mk-J_fEJKkogtr3NaSVolWh5Rv5DcehuReLphIyvY7VG-iypvHqGc7knVZXv2FWJaFed1bQaA2jkGFbmEmQs4DydnmwOdlA4Df_o8SjCPEW8Bq54SWJz6uNI8L9p0_e8ZllW1gaF0zCOuoQn_XdEm8jUW6o1h2N1o9YLFAbj3-w0rj5BWqOfl6ysEkeY0DXe7xm8Ycmd7PpC98edYtnkUgs5Kfs7H_4bn0E4TSP_1XS8=; expires=Thu, 01 Jan 1970 00:00:00 GMT; path=/signin-google; secure; samesite=none; httponly set-cookie: Identity.External=CfDJ8L16rcoVAjJMt5TusKivrFecjgIwf31r3SrfRXOwnrnUe9wLJ38mJtd8HAzKOc_2rsgrFsLc6lAnKcHa5pYmafjUFaKJcmBZt2ys2JSUl3qFKVqSTZaSpS4E21Tt70yaz2T9AzSzDOhBv2hIBrtj1Y9c1Ximra9zPlTXLI1PuUNPl7ZQSTzQfep7-hhr7Fj0aCz4Qq14UvjHBaU8QjGd_XACdODJc4aX1PlT8_JgXDl5ob9RtDxmta1egpT6EkFOFwBC1Ngb0UiY5Ilz-c1R3255SDIxa6PVVNjy55ccFo0Hpfik9JWj6oTgGApt6koNrxbFOTK5sLAt5S7lR1FbL8Jkq_DOpmTzfSgEPVV7PIfW6AK4zZav6qawqgTn01mYRM-NO8f_gcrhTBfQpEbrYAFqVTu-Svo6ZkKTI-iRMe57arPz8e8nUvRnqplAvR8MbRKvXNAHGdRcN-JDvzxZQHxm7I0u-gX0MUm1CezYDhbaENhosyZrcxZhNZ9u7XdVgEtAlPe_KX1Vdn0zHmgHadcW_q8l1yvVzMTigryWr8-6wvGyeCPVhGGfRH6vIqkF9hGgTNXLpIso3xESmRJveCXbYWZfZ9Tm2w51T4ToCwOc5ymNmYkUBvQYX0HCpKkwOIlCn1JLGWy7q6WcN1VRcZHImvX-TA3R9aiIqRwbuAe3Ju4Qop8oAGoXW1FxFPAGIQRytj_m7uPBdngitb5bZgcFKhXr5vgxngWrE2Ns55ItXu85aGHi8DymfUmuH-Wtsd6haLpq39E08D_zudyY11hbt2_raO9OPNkWV8R_QCYR4eij7sp1OTG_Jq1tkdsQRTe0DfmxE9-reYYw8hK_sF4Hb8GU8QO0Zrn0Ia9UrF6tqW-LNhO5Id02Jmxl2Yn6yOipNwO9FbRDmLy9O4LumZWZw1miLS-XJ9KxPRPcR3x_wCONtKrxfWZzEqYs7T34CSQB71okiKFwfMu_vUNJl44TUsmgZ9QEtgIxlHcSeIHPc7Y2FYU3ZEM3gBmekbGGywzQ48TxNedsHo9PJ_We5BHbyBQK5XvVJ8r22VBOYlZHaNbcOt3WS7KxJp-uHmlfCY7NuC5d27IqVN0Ir5sLDbnr6KSblI2y82faQXoXUQz-HQlTly4O6lEJc1YbEAcfgHWieye-3A9wbjEt6N7HISTjcWWNgJIk_IfJcQC2lP95wP5VHZgwKQrIitkRCYaUxwfTEqUzVZq34chEd9Y8eF1A-LAJjHpT4nsovrRuD1qof63XmJ-dqwdAEeC41CLRDgWRZRdAsFj2xEZkTIDu8IQZugOfxucgYBH8TxWIssU7YLtR4CVQoDOCmT-EIodwEPAY0Bi_mMHh6O0EhrGyXsvEFFm_M-A-6uLZgrL8jjU2DNUG9ZhFSIUncDZ1_9iEvuZfv8OMUCWayEmmml8-SXeElPvZTDk4O2Els11JRP7q1jTHZ-5vakdbRDql68HiHryY8xEsDN_4LW_blaHCwKV32npzrB_vnETmj0Ii7M1QytZV3Nfdr8R-bfYg6wF46z5bm-9TwWkX4_pwJ5dp116kvE5LkADo-NIrNo33EJCsp-iYYhmUWRVL5Rh7nszVHr-1fj2TramHcKRn7_oi4djcFIHl0w-i4dF3kRPy-dgE2nKHsSM9Rpyjn-whNSQfPM7AOxIYTbllejAmGVJxUU83TIrQiOeNxml6e5mlp3GrVmC5qtBmnUA4hdrpyGgWFzXx5uBCtLN2z39HrrKvIKD25naRPqu6mRHI5QRqSFL9mpmv4Cseqk975MnuUPtjI-8BhpGoavN_4PALb_VeM-JTh3TW0W40S6BOi3c97ZojY_J34EEO-WSk5zNw1WHS233etkQ5qCsjtae2f3_dC9Gi_SI9CgFgEnz6LjqqaXOisgVGJTGC-D4V-Ms0teHoEklFYrvhUMZiCmxGdW8FwrZF6bG9uAKX1iRvg_557SIczHx24v9zXZL33dN0m11THnnD_0yVAc6ac9uiXBWmI0eZ-0jxk0Rg2dZsZ75RS-Mj8CNQZPfPfvfQj9LpZh8uyS3c3IPJVLzLyVpmEC0co-8gty45-qFU-cjoyiPI3v9hH9HWcO6bcdp7w1dgj-nQxPCj0Sb8MkPbW2cvl2y9Hwr5MmAjEp8gnCtLlfvm7hbbzwntkQURVP2V7ir83ZjNfObKcXN7SPbcsf88yTNfnilrLyF-GfaufOiqz5PEEy33k2geQMRbjbjdzpEarQjLZnQGLZ5hupMmbr7UdME2e1PCZPBp-KWZP4qA8uW3gCsmSqhBQgo4fcmjFocPPyt1IEnhmJqfI8c32_CvCK8Df98f183xwuBBVJr23MN1rVYf4avXEnrTfrs6I6CJfy8ZN5upxcUqpMxTQiMcO1lVycx8WYvcZTl1VXc9CipLiE7t92rqLeoyUQENWSvryTIL5JDtEexJTZ5fGDyLc1NCq5JidkG-oH32aPtSwm1hEQa8MsWfK64N_8f5YA5KVQuen-d-IQil_u7LVNS5BfIjEuhydZqRdoEi-8IKoraDCFzs6prdaXzqB3rQJwa2cOnuCjQNT8xz2kvRBCq75rt7I-srPSYNA57W9SwP5dn8GR2AxUI8CcO7zCfuBwO4FbcEgBtxCWaDmhAVnZd91MH0sslMo4hqNI8PQnNNvtGDXuHsM68tETZflcEW1EEH6Ylr2LWCDGlEZEK6PI1QyYI8TteMw0Rv6dxHbJqhYtvmb3Mx9FIf48NCEWNOnHK6DLt4R6gf-DjZp8pe8z1yMQ8sUccE_4R53irREwFrwij8BUHm1ncDzAVBK8uaGoHskM4oeLjKn8xCEW5pOuOBzXryrt9d0-3ymmLXUsI8du2WB0LmsMOJ3IM4f7zi9P-TjWTt4bdJ7YAmF5lp_zU0ltvMexKB9aTj-b7gfoEd2nBL3rd4E_VCqh_WosaI_t27CulyuuWCsGEFChO1vC0ONeyKHcjJSBv7E3kUYoYvo664SGC8OBZe69VCxFSPOhersp8XS74F6VtTzj7B8xvht8IsuOmZ7Ce5dj6ZuacUarAVppJBzeFfE2HrDOKLQGZ9Irq9berzA9zmQlHFDGNbj3GkZ5x8RmAc-pn1w_GAHeIVzULfNZ8Gxc-3AkqwlqkbDVZEuUL10zgtjh9vezwhiAvllcKS86AunCdb6UKegoSKPHmAWHQBLNwVYBjs5lDBOlxpzN41UhHAAlKmBfBS4QDoejTZN8yZGWS7irf-9VmBDiI1oSh6UbajAXLVmKJ6cE2zqXBjrl_YEpqyPVIIENYDuaa2v4f_t-ErgUQb3QN7jOKVF_QhOru3NQyWcZcBPEUlStTnayT1V0LK_2yjApvG9tLGgpZVGdsza9sbMexUQjRu3phFS1maVP3nV3Ee5D2mWWCzOwqjRvgOObzR4XxY8-9hIDI7iho3mNhpoiNubvH4s2ocmfImHmZqRBuLB-pNZ4dSAAnPIBDLpVWLmLLnxp5zIs2phfI7_GrHAz3LUj1ODVT2lyBjQYI_bNgpBHzfKhNfMBnHm4HpC_d7v7tD4X1cD_dOCBu3VQIs77u60sjOemot7g_F__P9GdrJQC60jZyWnW7IOE8V1Kg1ZtRISRBGavq3kj37H1qLHqEE6_R4tIgUvQiWb-9BoegnTKaX6_Cpq4L_OFGiegOr7EalAf2KcDm1AL15FEcHuYlHe5YeW-XGaHQdG7Ovt2owKAO1Kg; path=/; secure; samesite=lax; httponly status: 302 x-correlation-id: d5a7b7fe-c5e9-4ab0-b6ae-0cb3ba891827 FORM-DATA state=CfDJ8L16rcoVAjJMt5TusKivrFf2GEKRNyWN6Am9QMgqgxz17T9ThyCs-KUxjv1eUpwqojsNxikmLmsC5QXNBURb4gj8IzcQ1lYVPx4iP20oEk3bO2ZQgibTGRDlkBOAp6S7fSCUOPi1lwI1a5H9JjmewSY_4lpEftzbPKn9PCIbHuIkEcbh1UIDkptbanBb-ALt40tmtX8_UzgR7ZQJLLzzNp7POLvKSSXR8qugXU8CefSHCJGGNFM4GnNEVkRVW4oIhm2afXjlLJ13D90yaa8XzVPz_gHAiCgrCM4LVaDqTBfjHWZDiDAyDl66H_XGTaYjyNosKGStsZR3pF-23O0BQrmaXriT6hPTxKQnimMytT3DREAG0G6otxk_mBBMbQziepTeIqNWAMePCEOXZ5Ha367R343Kzyill8D_YNt82uTcFmf9xe-pAT7Mn9Q6P-p0Vsvs5WWN8dQyXposYpUThN-Ub6IoK4WtnNI2gYKBMR_S&code=REDACTED&scope=profile+openid+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.profile&authuser=0&prompt=none ``` ```ExternalLoginModel.OnGetCallbackAsync``` gets called but then I get a ```null``` when calling ```GetExternalLoginInfoAsync ``` I recall that some cookies get cleared during this exchange, but why would simply hosting my app in an iFrame result in different behavior?
non_test
getexternallogininfoasync is null when called from externalloginmodel ongetcallbackasync iframes involved goal perform a silent login using a hidden iframe to google result i can achieve this by passing prompt none to google which then doesn t x frame options deny me all good when i run the app normally and i route to and my plugin page has my hidden iframe signin i ve heard nothing yet form style display none id loginform target guestframe action identity account externallogin returnurl amp errorurl amp prompt none amp provider google method post your browser does not support inline frames ok now put the entire app in an iframe and try to do a silent login under that context iframe id guestframe name guestframe src the above iframe is hosted in i am using as my wildcard dns my signin google gets hit and it looks like there is a good code in there general request url request method post status code remote address referrer policy no referrer when downgrade response headers cache control no cache content length date fri aug gmt expires thu jan gmt location identity account externallogin returnurl handler callback pragma no cache server kestrel set cookie aspnetcore correlation google j ztc jyvsnaf iq expires thu jan gmt path signin google secure samesite none httponly set cookie aspnetcore openidconnect nonce j expires thu jan gmt path signin google secure samesite none httponly set cookie identity external gcrhtbfqpebryafqvtu m a vem qfu d iqil vcqh wosai t f path secure samesite lax httponly status x correlation id form data state s code redacted scope profile openid https externalloginmodel ongetcallbackasync gets called but then i get a null when calling getexternallogininfoasync i recall that some cookies get cleared during this exchange but why would simply hosting my app in an iframe result in different behavior
0
117,183
4,712,108,175
IssuesEvent
2016-10-14 15:44:44
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Move the visibility icon (the eye) to the side of the tools for the layer in the layer tree
enhancement Priority: High task
Right now the eye icon that allows users to change the visibility of the layer is in a position which is strange, see below: ![clipboard021](https://cloud.githubusercontent.com/assets/693880/19234913/d4e35884-8eee-11e6-8b5f-c2d995dfa400.jpg) I would rather have something like this: ![clipboard031](https://cloud.githubusercontent.com/assets/693880/19234936/03b85f10-8eef-11e6-96bc-2302aa08c3ce.jpg) which can be taken from the SIRA product we are doing.
1.0
Move the visibility icon (the eye) to the side of the tools for the layer in the layer tree - Right now the eye icon that allows users to change the visibility of the layer is in a position which is strange, see below: ![clipboard021](https://cloud.githubusercontent.com/assets/693880/19234913/d4e35884-8eee-11e6-8b5f-c2d995dfa400.jpg) I would rather have something like this: ![clipboard031](https://cloud.githubusercontent.com/assets/693880/19234936/03b85f10-8eef-11e6-96bc-2302aa08c3ce.jpg) which can be taken from the SIRA product we are doing.
non_test
move the visibility icon the eye to the side of the tools for the layer in the layer tree right now the eye icon that allows users to change the visibility of the layer is in a position which is strange see below i would rather have something like this which can be taken from the sira product we are doing
0
160,424
20,100,272,398
IssuesEvent
2022-02-07 02:35:53
dwanprac/django.nv
https://api.github.com/repos/dwanprac/django.nv
opened
grunt-npm-install-0.3.1.tgz: 28 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>grunt-npm-install-0.3.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-3918](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | json-schema-0.2.3.tgz | Transitive | N/A | ❌ | | [CVE-2018-16492](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | extend-3.0.0.tgz | Transitive | N/A | ❌ | | [CVE-2021-23807](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | jsonpointer-4.0.0.tgz | Transitive | N/A | ❌ | | [WS-2020-0344](https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [WS-2020-0345](https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.2 | jsonpointer-4.0.0.tgz | Transitive | N/A | ❌ | | [CVE-2019-16776](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2021-32804](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | tar-2.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2018-20834](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tar-2.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2017-15010](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tough-cookie-2.3.1.tgz | Transitive | N/A | ❌ | | [WS-2020-0180](https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | npm-user-validate-0.1.5.tgz | Transitive | N/A | ❌ | | [WS-2018-0069](https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [CVE-2017-1000048](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | qs-6.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2020-7754](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7754) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | npm-user-validate-0.1.5.tgz | Transitive | N/A | ❌ | | [CVE-2018-3737](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | sshpk-1.10.1.tgz | Transitive | N/A | ❌ | | [CVE-2017-18077](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18077) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | brace-expansion-1.1.6.tgz | Transitive | N/A | ❌ | | [WS-2020-0342](https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [CVE-2019-13173](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | fstream-1.0.10.tgz | Transitive | N/A | ❌ | | [CVE-2020-7788](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | ini-1.3.4.tgz | Transitive | N/A | ❌ | | [CVE-2018-21270](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | stringstream-0.0.5.tgz | Transitive | N/A | ❌ | | [CVE-2020-8244](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | bl-1.1.2.tgz | Transitive | N/A | ❌ | | [CVE-2019-16775](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16775) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2019-16777](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16777) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2017-16032](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16032) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | brace-expansion-1.1.6.tgz | Transitive | N/A | ❌ | | [CVE-2021-23362](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | hosted-git-info-2.1.5.tgz | Transitive | N/A | ❌ | | [CVE-2018-1107](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [WS-2018-0103](https://hackerone.com/reports/321670) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | stringstream-0.0.5.tgz | Transitive | N/A | ❌ | | [CVE-2020-15095](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15095) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.4 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2017-18869](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18869) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.5 | chownr-1.0.1.tgz | Transitive | N/A | ❌ | ## Details > Partial details (11 vulnerabilities) are displayed below due to a content size limitation in GitHub. To view information on the remaining vulnerabilities, navigate to the WhiteSource Application.<br> <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3918</summary> ### Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - http-signature-1.1.1.tgz - jsprim-1.3.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-16492</summary> ### Vulnerable Library - <b>extend-3.0.0.tgz</b></p> <p>Port of jQuery.extend for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/extend/-/extend-3.0.0.tgz">https://registry.npmjs.org/extend/-/extend-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/extend/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **extend-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A prototype pollution vulnerability was found in module extend <2.0.2, ~<3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492>CVE-2018-16492</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/381185">https://hackerone.com/reports/381185</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: extend - v3.0.2,v2.0.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23807</summary> ### Vulnerable Library - <b>jsonpointer-4.0.0.tgz</b></p> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.15.0.tgz - :x: **jsonpointer-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays. <p>Publish Date: 2021-11-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807>CVE-2021-23807</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807</a></p> <p>Release Date: 2021-11-03</p> <p>Fix Resolution: jsonpointer - 5.0.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0344</summary> ### Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function. <p>Publish Date: 2020-06-09 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-09</p> <p>Fix Resolution: is-my-json-valid - 2.20.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0345</summary> ### Vulnerable Library - <b>jsonpointer-4.0.0.tgz</b></p> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.15.0.tgz - :x: **jsonpointer-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function. <p>Publish Date: 2020-07-03 <p>URL: <a href=https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34>WS-2020-0345</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/janl/node-jsonpointer/releases/tag/v4.1.0">https://github.com/janl/node-jsonpointer/releases/tag/v4.1.0</a></p> <p>Release Date: 2020-07-03</p> <p>Fix Resolution: jsonpointer - 4.1.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-16776</summary> ### Vulnerable Library - <b>npm-3.10.10.tgz</b></p> <p>a package manager for JavaScript</p> <p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-3.10.10.tgz">https://registry.npmjs.org/npm/-/npm-3.10.10.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - :x: **npm-3.10.10.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option. <p>Publish Date: 2019-12-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p> <p>Release Date: 2020-10-07</p> <p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-32804</summary> ### Vulnerable Library - <b>tar-2.2.1.tgz</b></p> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-20834</summary> ### Vulnerable Library - <b>tar-2.2.1.tgz</b></p> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A vulnerability was found in node-tar before version 4.4.2 (excluding version 2.2.2). An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. A patch has been applied to node-tar v2.2.2). <p>Publish Date: 2019-04-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834>CVE-2018-20834</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 2.2.2,4.4.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-15010</summary> ### Vulnerable Library - <b>tough-cookie-2.3.1.tgz</b></p> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.3.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.3.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **tough-cookie-2.3.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A ReDoS (regular expression denial of service) flaw was found in the tough-cookie module before 2.3.3 for Node.js. An attacker that is able to make an HTTP request using a specially crafted cookie may cause the application to consume an excessive amount of CPU. <p>Publish Date: 2017-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010>CVE-2017-15010</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-15010">https://nvd.nist.gov/vuln/detail/CVE-2017-15010</a></p> <p>Release Date: 2017-10-04</p> <p>Fix Resolution: 2.3.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0180</summary> ### Vulnerable Library - <b>npm-user-validate-0.1.5.tgz</b></p> <p>User validations for npm</p> <p>Library home page: <a href="https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz">https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/npm-user-validate/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **npm-user-validate-0.1.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user's email took exponentially longer to process input strings that begin with the '@' character. <p>Publish Date: 2020-10-16 <p>URL: <a href=https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e>WS-2020-0180</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-xgh6-85xh-479p">https://github.com/advisories/GHSA-xgh6-85xh-479p</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: 1.0.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0069</summary> ### Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-02-14</p> <p>Fix Resolution: 1.4.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;jsprim:1.3.1;json-schema:0.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"json-schema - 0.4.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"extend","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;extend:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"extend - v3.0.2,v2.0.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-16492","vulnerabilityDetails":"A prototype pollution vulnerability was found in module extend \u003c2.0.2, ~\u003c3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jsonpointer","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0;jsonpointer:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jsonpointer - 5.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23807","vulnerabilityDetails":"This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jsonpointer","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0;jsonpointer:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jsonpointer - 4.1.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0345","vulnerabilityDetails":"Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function.","vulnerabilityUrl":"https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16776","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"2.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;tar:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"2.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;tar:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.2,4.4.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-20834","vulnerabilityDetails":"A vulnerability was found in node-tar before version 4.4.2 (excluding version 2.2.2). An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. A patch has been applied to node-tar v2.2.2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tough-cookie","packageVersion":"2.3.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;tough-cookie:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-15010","vulnerabilityDetails":"A ReDoS (regular expression denial of service) flaw was found in the tough-cookie module before 2.3.3 for Node.js. An attacker that is able to make an HTTP request using a specially crafted cookie may cause the application to consume an excessive amount of CPU.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm-user-validate","packageVersion":"0.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;npm-user-validate:0.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0180","vulnerabilityDetails":"The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user\u0027s email took exponentially longer to process input strings that begin with the \u0027@\u0027 character.","vulnerabilityUrl":"https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2018-0069","vulnerabilityDetails":"Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm-user-validate","packageVersion":"0.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;npm-user-validate:0.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7754","vulnerabilityDetails":"This affects the package npm-user-validate before 1.0.1. The regex that validates user emails took exponentially longer to process long input strings beginning with @ characters.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7754","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sshpk","packageVersion":"1.10.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;sshpk:1.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-3737","vulnerabilityDetails":"sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"brace-expansion","packageVersion":"1.1.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;init-package-json:1.9.4;glob:6.0.4;minimatch:3.0.3;brace-expansion:1.1.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.1.7","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-18077","vulnerabilityDetails":"index.js in brace-expansion before 1.1.7 is vulnerable to Regular Expression Denial of Service (ReDoS) attacks, as demonstrated by an expand argument containing many comma characters.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18077","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0342","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"fstream","packageVersion":"1.0.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;fstream:1.0.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.12","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-13173","vulnerabilityDetails":"fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system\u0027s file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ini","packageVersion":"1.3.4","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;ini:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.3.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7788","vulnerabilityDetails":"This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"stringstream","packageVersion":"0.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;stringstream:0.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.0.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-21270","vulnerabilityDetails":"Versions less than 0.0.6 of the Node.js stringstream module are vulnerable to an out-of-bounds read because of allocation of uninitialized buffers when a number is passed in the input stream (when using Node.js 4.x).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bl","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;bl:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bl - 1.2.3,2.2.1,3.0.1,4.0.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-8244","vulnerabilityDetails":"A buffer over-read vulnerability exists in bl \u003c4.0.3, \u003c3.0.1, \u003c2.2.1, and \u003c1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16775","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It is possible for packages to create symlinks to files outside of thenode_modules folder through the bin field upon installation. A properly constructed entry in the package.json bin field would allow a package publisher to create a symlink pointing to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16775","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16777","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.4 are vulnerable to an Arbitrary File Overwrite. It fails to prevent existing globally-installed binaries to be overwritten by other package installations. For example, if a package was installed globally and created a serve binary, any subsequent installs of packages that also create a serve binary would overwrite the previous serve binary. This behavior is still allowed in local installations and also through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16777","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"brace-expansion","packageVersion":"1.1.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;init-package-json:1.9.4;glob:6.0.4;minimatch:3.0.3;brace-expansion:1.1.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.1.7","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16032","vulnerabilityDetails":"brace-expansion before 1.1.7 are vulnerable to a regular expression denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16032","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"Required","AV":"Local","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;hosted-git-info:2.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 1.4.1, 2.17.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1107","vulnerabilityDetails":"It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"stringstream","packageVersion":"0.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;stringstream:0.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.0.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2018-0103","vulnerabilityDetails":"All versions of stringstream are vulnerable to out-of-bounds read as it allocates uninitialized Buffers when number is passed in input stream on Node.js 4.x and below.","vulnerabilityUrl":"https://hackerone.com/reports/321670","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.14.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-15095","vulnerabilityDetails":"Versions of the npm CLI prior to 6.14.6 are vulnerable to an information exposure vulnerability through log files. The CLI supports URLs like \"\u003cprotocol\u003e://[\u003cuser\u003e[:\u003cpassword\u003e]@]\u003chostname\u003e[:\u003cport\u003e][:][/]\u003cpath\u003e\". The password value is not redacted and is printed to stdout and also to any generated log files.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15095","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"Required","AV":"Local","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"chownr","packageVersion":"1.0.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;chownr:1.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.1.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-18869","vulnerabilityDetails":"A TOCTOU issue in the chownr package before 1.1.0 for Node.js 10.10 could allow a local attacker to trick it into descending into unintended directories via symlink attacks.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18869","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}]</REMEDIATE> -->
True
grunt-npm-install-0.3.1.tgz: 28 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>grunt-npm-install-0.3.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-3918](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | json-schema-0.2.3.tgz | Transitive | N/A | ❌ | | [CVE-2018-16492](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | extend-3.0.0.tgz | Transitive | N/A | ❌ | | [CVE-2021-23807](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | jsonpointer-4.0.0.tgz | Transitive | N/A | ❌ | | [WS-2020-0344](https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [WS-2020-0345](https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.2 | jsonpointer-4.0.0.tgz | Transitive | N/A | ❌ | | [CVE-2019-16776](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2021-32804](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | tar-2.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2018-20834](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tar-2.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2017-15010](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tough-cookie-2.3.1.tgz | Transitive | N/A | ❌ | | [WS-2020-0180](https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | npm-user-validate-0.1.5.tgz | Transitive | N/A | ❌ | | [WS-2018-0069](https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [CVE-2017-1000048](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | qs-6.2.1.tgz | Transitive | N/A | ❌ | | [CVE-2020-7754](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7754) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | npm-user-validate-0.1.5.tgz | Transitive | N/A | ❌ | | [CVE-2018-3737](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | sshpk-1.10.1.tgz | Transitive | N/A | ❌ | | [CVE-2017-18077](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18077) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | brace-expansion-1.1.6.tgz | Transitive | N/A | ❌ | | [WS-2020-0342](https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [CVE-2019-13173](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | fstream-1.0.10.tgz | Transitive | N/A | ❌ | | [CVE-2020-7788](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | ini-1.3.4.tgz | Transitive | N/A | ❌ | | [CVE-2018-21270](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | stringstream-0.0.5.tgz | Transitive | N/A | ❌ | | [CVE-2020-8244](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | bl-1.1.2.tgz | Transitive | N/A | ❌ | | [CVE-2019-16775](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16775) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2019-16777](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16777) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2017-16032](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16032) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | brace-expansion-1.1.6.tgz | Transitive | N/A | ❌ | | [CVE-2021-23362](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | hosted-git-info-2.1.5.tgz | Transitive | N/A | ❌ | | [CVE-2018-1107](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.15.0.tgz | Transitive | N/A | ❌ | | [WS-2018-0103](https://hackerone.com/reports/321670) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | stringstream-0.0.5.tgz | Transitive | N/A | ❌ | | [CVE-2020-15095](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15095) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.4 | npm-3.10.10.tgz | Transitive | N/A | ❌ | | [CVE-2017-18869](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18869) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.5 | chownr-1.0.1.tgz | Transitive | N/A | ❌ | ## Details > Partial details (11 vulnerabilities) are displayed below due to a content size limitation in GitHub. To view information on the remaining vulnerabilities, navigate to the WhiteSource Application.<br> <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3918</summary> ### Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - http-signature-1.1.1.tgz - jsprim-1.3.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-16492</summary> ### Vulnerable Library - <b>extend-3.0.0.tgz</b></p> <p>Port of jQuery.extend for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/extend/-/extend-3.0.0.tgz">https://registry.npmjs.org/extend/-/extend-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/extend/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **extend-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A prototype pollution vulnerability was found in module extend <2.0.2, ~<3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492>CVE-2018-16492</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/381185">https://hackerone.com/reports/381185</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: extend - v3.0.2,v2.0.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23807</summary> ### Vulnerable Library - <b>jsonpointer-4.0.0.tgz</b></p> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.15.0.tgz - :x: **jsonpointer-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays. <p>Publish Date: 2021-11-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807>CVE-2021-23807</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807</a></p> <p>Release Date: 2021-11-03</p> <p>Fix Resolution: jsonpointer - 5.0.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0344</summary> ### Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function. <p>Publish Date: 2020-06-09 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-09</p> <p>Fix Resolution: is-my-json-valid - 2.20.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0345</summary> ### Vulnerable Library - <b>jsonpointer-4.0.0.tgz</b></p> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.15.0.tgz - :x: **jsonpointer-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function. <p>Publish Date: 2020-07-03 <p>URL: <a href=https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34>WS-2020-0345</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/janl/node-jsonpointer/releases/tag/v4.1.0">https://github.com/janl/node-jsonpointer/releases/tag/v4.1.0</a></p> <p>Release Date: 2020-07-03</p> <p>Fix Resolution: jsonpointer - 4.1.0</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-16776</summary> ### Vulnerable Library - <b>npm-3.10.10.tgz</b></p> <p>a package manager for JavaScript</p> <p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-3.10.10.tgz">https://registry.npmjs.org/npm/-/npm-3.10.10.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - :x: **npm-3.10.10.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option. <p>Publish Date: 2019-12-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p> <p>Release Date: 2020-10-07</p> <p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-32804</summary> ### Vulnerable Library - <b>tar-2.2.1.tgz</b></p> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-20834</summary> ### Vulnerable Library - <b>tar-2.2.1.tgz</b></p> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A vulnerability was found in node-tar before version 4.4.2 (excluding version 2.2.2). An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. A patch has been applied to node-tar v2.2.2). <p>Publish Date: 2019-04-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834>CVE-2018-20834</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 2.2.2,4.4.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-15010</summary> ### Vulnerable Library - <b>tough-cookie-2.3.1.tgz</b></p> <p>RFC6265 Cookies and Cookie Jar for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.3.1.tgz">https://registry.npmjs.org/tough-cookie/-/tough-cookie-2.3.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/tough-cookie/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **tough-cookie-2.3.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A ReDoS (regular expression denial of service) flaw was found in the tough-cookie module before 2.3.3 for Node.js. An attacker that is able to make an HTTP request using a specially crafted cookie may cause the application to consume an excessive amount of CPU. <p>Publish Date: 2017-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010>CVE-2017-15010</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-15010">https://nvd.nist.gov/vuln/detail/CVE-2017-15010</a></p> <p>Release Date: 2017-10-04</p> <p>Fix Resolution: 2.3.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0180</summary> ### Vulnerable Library - <b>npm-user-validate-0.1.5.tgz</b></p> <p>User validations for npm</p> <p>Library home page: <a href="https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz">https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/npm-user-validate/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **npm-user-validate-0.1.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user's email took exponentially longer to process input strings that begin with the '@' character. <p>Publish Date: 2020-10-16 <p>URL: <a href=https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e>WS-2020-0180</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-xgh6-85xh-479p">https://github.com/advisories/GHSA-xgh6-85xh-479p</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: 1.0.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0069</summary> ### Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dwanprac/django.nv/commit/b5ec9bd74b4b27b2af49cf824e527f391e9437b0">b5ec9bd74b4b27b2af49cf824e527f391e9437b0</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-02-14</p> <p>Fix Resolution: 1.4.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;jsprim:1.3.1;json-schema:0.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"json-schema - 0.4.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"extend","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;extend:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"extend - v3.0.2,v2.0.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-16492","vulnerabilityDetails":"A prototype pollution vulnerability was found in module extend \u003c2.0.2, ~\u003c3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jsonpointer","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0;jsonpointer:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jsonpointer - 5.0.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23807","vulnerabilityDetails":"This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"jsonpointer","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0;jsonpointer:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jsonpointer - 4.1.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0345","vulnerabilityDetails":"Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function.","vulnerabilityUrl":"https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16776","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"2.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;tar:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"2.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;tar:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.2,4.4.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-20834","vulnerabilityDetails":"A vulnerability was found in node-tar before version 4.4.2 (excluding version 2.2.2). An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. A patch has been applied to node-tar v2.2.2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tough-cookie","packageVersion":"2.3.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;tough-cookie:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-15010","vulnerabilityDetails":"A ReDoS (regular expression denial of service) flaw was found in the tough-cookie module before 2.3.3 for Node.js. An attacker that is able to make an HTTP request using a specially crafted cookie may cause the application to consume an excessive amount of CPU.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm-user-validate","packageVersion":"0.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;npm-user-validate:0.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0180","vulnerabilityDetails":"The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user\u0027s email took exponentially longer to process input strings that begin with the \u0027@\u0027 character.","vulnerabilityUrl":"https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2018-0069","vulnerabilityDetails":"Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm-user-validate","packageVersion":"0.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;npm-user-validate:0.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7754","vulnerabilityDetails":"This affects the package npm-user-validate before 1.0.1. The regex that validates user emails took exponentially longer to process long input strings beginning with @ characters.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7754","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sshpk","packageVersion":"1.10.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;sshpk:1.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-3737","vulnerabilityDetails":"sshpk is vulnerable to ReDoS when parsing crafted invalid public keys.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"brace-expansion","packageVersion":"1.1.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;init-package-json:1.9.4;glob:6.0.4;minimatch:3.0.3;brace-expansion:1.1.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.1.7","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-18077","vulnerabilityDetails":"index.js in brace-expansion before 1.1.7 is vulnerable to Regular Expression Denial of Service (ReDoS) attacks, as demonstrated by an expand argument containing many comma characters.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18077","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0342","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"fstream","packageVersion":"1.0.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;fstream:1.0.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.12","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-13173","vulnerabilityDetails":"fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system\u0027s file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ini","packageVersion":"1.3.4","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;ini:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.3.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7788","vulnerabilityDetails":"This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"stringstream","packageVersion":"0.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;stringstream:0.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.0.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-21270","vulnerabilityDetails":"Versions less than 0.0.6 of the Node.js stringstream module are vulnerable to an out-of-bounds read because of allocation of uninitialized buffers when a number is passed in the input stream (when using Node.js 4.x).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bl","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;bl:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bl - 1.2.3,2.2.1,3.0.1,4.0.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-8244","vulnerabilityDetails":"A buffer over-read vulnerability exists in bl \u003c4.0.3, \u003c3.0.1, \u003c2.2.1, and \u003c1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16775","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It is possible for packages to create symlinks to files outside of thenode_modules folder through the bin field upon installation. A properly constructed entry in the package.json bin field would allow a package publisher to create a symlink pointing to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16775","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16777","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.4 are vulnerable to an Arbitrary File Overwrite. It fails to prevent existing globally-installed binaries to be overwritten by other package installations. For example, if a package was installed globally and created a serve binary, any subsequent installs of packages that also create a serve binary would overwrite the previous serve binary. This behavior is still allowed in local installations and also through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16777","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"brace-expansion","packageVersion":"1.1.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;init-package-json:1.9.4;glob:6.0.4;minimatch:3.0.3;brace-expansion:1.1.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.1.7","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16032","vulnerabilityDetails":"brace-expansion before 1.1.7 are vulnerable to a regular expression denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16032","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"Required","AV":"Local","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.1.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;hosted-git-info:2.1.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 1.4.1, 2.17.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1107","vulnerabilityDetails":"It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"stringstream","packageVersion":"0.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;stringstream:0.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.0.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2018-0103","vulnerabilityDetails":"All versions of stringstream are vulnerable to out-of-bounds read as it allocates uninitialized Buffers when number is passed in input stream on Node.js 4.x and below.","vulnerabilityUrl":"https://hackerone.com/reports/321670","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.14.6","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-15095","vulnerabilityDetails":"Versions of the npm CLI prior to 6.14.6 are vulnerable to an information exposure vulnerability through log files. The CLI supports URLs like \"\u003cprotocol\u003e://[\u003cuser\u003e[:\u003cpassword\u003e]@]\u003chostname\u003e[:\u003cport\u003e][:][/]\u003cpath\u003e\". The password value is not redacted and is printed to stdout and also to any generated log files.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15095","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"Required","AV":"Local","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"chownr","packageVersion":"1.0.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;chownr:1.0.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.1.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-18869","vulnerabilityDetails":"A TOCTOU issue in the chownr package before 1.1.0 for Node.js 10.10 could allow a local attacker to trick it into descending into unintended directories via symlink attacks.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18869","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}]</REMEDIATE> -->
non_test
grunt npm install tgz vulnerabilities highest severity is vulnerable library grunt npm install tgz path to dependency file package json path to vulnerable library node modules npm node modules request node modules har validator node modules is my json valid package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high json schema tgz transitive n a ❌ high extend tgz transitive n a ❌ high jsonpointer tgz transitive n a ❌ high is my json valid tgz transitive n a ❌ high jsonpointer tgz transitive n a ❌ high npm tgz transitive n a ❌ high tar tgz transitive n a ❌ high tar tgz transitive n a ❌ high tough cookie tgz transitive n a ❌ high npm user validate tgz transitive n a ❌ high is my json valid tgz transitive n a ❌ high qs tgz transitive n a ❌ high npm user validate tgz transitive n a ❌ high sshpk tgz transitive n a ❌ high brace expansion tgz transitive n a ❌ high is my json valid tgz transitive n a ❌ high fstream tgz transitive n a ❌ high ini tgz transitive n a ❌ medium stringstream tgz transitive n a ❌ medium bl tgz transitive n a ❌ medium npm tgz transitive n a ❌ medium npm tgz transitive n a ❌ medium brace expansion tgz transitive n a ❌ medium hosted git info tgz transitive n a ❌ medium is my json valid tgz transitive n a ❌ medium stringstream tgz transitive n a ❌ medium npm tgz transitive n a ❌ low chownr tgz transitive n a ❌ details partial details vulnerabilities are displayed below due to a content size limitation in github to view information on the remaining vulnerabilities navigate to the whitesource application cve vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules http signature node modules jsprim node modules json schema package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href found in base branch main vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema step up your open source security game with whitesource cve vulnerable library extend tgz port of jquery extend for node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules extend package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz x extend tgz vulnerable library found in head commit a href found in base branch main vulnerability details a prototype pollution vulnerability was found in module extend that allows an attacker to inject arbitrary properties onto object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution extend step up your open source security game with whitesource cve vulnerable library jsonpointer tgz simple json addressing library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules har validator node modules is my json valid node modules jsonpointer package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz is my json valid tgz x jsonpointer tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package jsonpointer before a type confusion vulnerability can lead to a bypass of a previous prototype pollution fix when the pointer components are arrays publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonpointer step up your open source security game with whitesource ws vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch main vulnerability details arbitrary code execution vulnerability was found in is my json valid before via the fromatname function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid step up your open source security game with whitesource ws vulnerable library jsonpointer tgz simple json addressing library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules har validator node modules is my json valid node modules jsonpointer package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz is my json valid tgz x jsonpointer tgz vulnerable library found in head commit a href found in base branch main vulnerability details prototype pollution vulnerability was found in jsonpointer before via the set function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonpointer step up your open source security game with whitesource cve vulnerable library npm tgz a package manager for javascript library home page a href path to dependency file package json path to vulnerable library node modules npm package json dependency hierarchy grunt npm install tgz root library x npm tgz vulnerable library found in head commit a href found in base branch main vulnerability details versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user s system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution npm yarn step up your open source security game with whitesource cve vulnerable library tar tgz tar for node library home page a href path to dependency file package json path to vulnerable library node modules npm node modules tar package json dependency hierarchy grunt npm install tgz root library npm tgz x tar tgz vulnerable library found in head commit a href found in base branch main vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource cve vulnerable library tar tgz tar for node library home page a href path to dependency file package json path to vulnerable library node modules npm node modules tar package json dependency hierarchy grunt npm install tgz root library npm tgz x tar tgz vulnerable library found in head commit a href found in base branch main vulnerability details a vulnerability was found in node tar before version excluding version an arbitrary file overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system in conjunction with a later plain file with the same name as the hardlink this plain file content replaces the existing file content a patch has been applied to node tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource cve vulnerable library tough cookie tgz cookies and cookie jar for node js library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules tough cookie package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz x tough cookie tgz vulnerable library found in head commit a href found in base branch main vulnerability details a redos regular expression denial of service flaw was found in the tough cookie module before for node js an attacker that is able to make an http request using a specially crafted cookie may cause the application to consume an excessive amount of cpu publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ws vulnerable library npm user validate tgz user validations for npm library home page a href path to dependency file package json path to vulnerable library node modules npm node modules npm user validate package json dependency hierarchy grunt npm install tgz root library npm tgz x npm user validate tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package npm user validate prior to version is vulnerable to redos the regex that validates a user s email took exponentially longer to process input strings that begin with the character publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource ws vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch main vulnerability details version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource istransitivedependency true dependencytree grunt npm install npm request http signature jsprim json schema isminimumfixversionavailable true minimumfixversion json schema isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails json schema is vulnerable to improperly controlled modification of object prototype attributes pollution vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request extend isminimumfixversionavailable true minimumfixversion extend isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a prototype pollution vulnerability was found in module extend that allows an attacker to inject arbitrary properties onto object prototype vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid jsonpointer isminimumfixversionavailable true minimumfixversion jsonpointer isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package jsonpointer before a type confusion vulnerability can lead to a bypass of a previous prototype pollution fix when the pointer components are arrays vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails arbitrary code execution vulnerability was found in is my json valid before via the fromatname function vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid jsonpointer isminimumfixversionavailable true minimumfixversion jsonpointer isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails prototype pollution vulnerability was found in jsonpointer before via the set function vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm isminimumfixversionavailable true minimumfixversion npm yarn isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm tar isminimumfixversionavailable true minimumfixversion tar isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm tar isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a vulnerability was found in node tar before version excluding version an arbitrary file overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system in conjunction with a later plain file with the same name as the hardlink this plain file content replaces the existing file content a patch has been applied to node tar vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request tough cookie isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a redos regular expression denial of service flaw was found in the tough cookie module before for node js an attacker that is able to make an http request using a specially crafted cookie may cause the application to consume an excessive amount of cpu vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm npm user validate isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails the package npm user validate prior to version is vulnerable to redos the regex that validates a user email took exponentially longer to process input strings that begin with the character vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request qs isminimumfixversionavailable true minimumfixversion qs isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the web framework using ljharb qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm npm user validate isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package npm user validate before the regex that validates user emails took exponentially longer to process long input strings beginning with characters vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request http signature sshpk isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails sshpk is vulnerable to redos when parsing crafted invalid public keys vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm init package json glob minimatch brace expansion isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails index js in brace expansion before is vulnerable to regular expression denial of service redos attacks as demonstrated by an expand argument containing many comma characters vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails regular expression denial of service redos vulnerability was found in is my json valid before via the style format vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm fstream isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fstream before is vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system file with the contents of the extracted file the fstream dirwriter function is vulnerable vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm ini isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package ini before if an attacker submits a malicious ini file to an application that parses it with ini parse they will pollute the prototype on the application this can be exploited further depending on the context vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request stringstream isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions less than of the node js stringstream module are vulnerable to an out of bounds read because of allocation of uninitialized buffers when a number is passed in the input stream when using node js x vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request bl isminimumfixversionavailable true minimumfixversion bl isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a buffer over read vulnerability exists in bl and which could allow an attacker to supply user input even typed that if it ends up in consume argument and can become negative the bufferlist state can be corrupted tricking it into exposing uninitialized memory via regular slice calls vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm isminimumfixversionavailable true minimumfixversion npm yarn isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of the npm cli prior to are vulnerable to an arbitrary file write it is possible for packages to create symlinks to files outside of thenode modules folder through the bin field upon installation a properly constructed entry in the package json bin field would allow a package publisher to create a symlink pointing to arbitrary files on a user system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm isminimumfixversionavailable true minimumfixversion npm isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of the npm cli prior to are vulnerable to an arbitrary file overwrite it fails to prevent existing globally installed binaries to be overwritten by other package installations for example if a package was installed globally and created a serve binary any subsequent installs of packages that also create a serve binary would overwrite the previous serve binary this behavior is still allowed in local installations and also through install scripts this vulnerability bypasses a user using the ignore scripts install option vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm init package json glob minimatch brace expansion isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails brace expansion before are vulnerable to a regular expression denial of service vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm hosted git info isminimumfixversionavailable true minimumfixversion hosted git info isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails it was discovered that the is my json valid javascript library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm request stringstream isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails all versions of stringstream are vulnerable to out of bounds read as it allocates uninitialized buffers when number is passed in input stream on node js x and below vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm isminimumfixversionavailable true minimumfixversion npm isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of the npm cli prior to are vulnerable to an information exposure vulnerability through log files the cli supports urls like the password value is not redacted and is printed to stdout and also to any generated log files vulnerabilityurl istransitivedependency true dependencytree grunt npm install npm chownr isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a toctou issue in the chownr package before for node js could allow a local attacker to trick it into descending into unintended directories via symlink attacks vulnerabilityurl
0
202,185
15,265,649,195
IssuesEvent
2021-02-22 07:40:24
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
kv/kvserver: TestStrictGCEnforcement failed
C-test-failure O-robot branch-master
[(kv/kvserver).TestStrictGCEnforcement failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2697767&tab=buildLog) on [master@8b6f3c84cc256debeeb4d4055c5f0d5c9a481213](https://github.com/cockroachdb/cockroach/commits/8b6f3c84cc256debeeb4d4055c5f0d5c9a481213): ``` === RUN TestStrictGCEnforcement test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestStrictGCEnforcement865385043 test_log_scope.go:74: use -show-logs to present logs inline === CONT TestStrictGCEnforcement client_replica_test.go:3277: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestStrictGCEnforcement865385043 --- FAIL: TestStrictGCEnforcement (0.58s) === RUN TestStrictGCEnforcement/protected_timestamps_are_respected client_replica_test.go:3275: Error Trace: client_replica_test.go:3148 client_replica_test.go:3275 Error: Received unexpected error: batch timestamp 1613803810.475835111,0 must be after replica GC threshold 1613803819.542992787,0 Test: TestStrictGCEnforcement/protected_timestamps_are_respected E210220 06:50:20.549634 490766 kv/kvclient/kvcoord/txn_interceptor_committer.go:449 [-] 1 making txn commit explicit failed for "sql txn" meta={id=e4613806 pri=0.00440269 epo=0 ts=1613803820.543560449,0 min=1613803820.543560449,0 seq=6} lock=true stat=STAGING rts=1613803820.543560449,0 wto=false gul=1613803821.043560449,0 ifw=5: result is ambiguous (server shutdown) --- FAIL: TestStrictGCEnforcement/protected_timestamps_are_respected (0.02s) ``` <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-parallel=4 ``` make stressrace TESTS=TestStrictGCEnforcement PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStrictGCEnforcement.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
kv/kvserver: TestStrictGCEnforcement failed - [(kv/kvserver).TestStrictGCEnforcement failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2697767&tab=buildLog) on [master@8b6f3c84cc256debeeb4d4055c5f0d5c9a481213](https://github.com/cockroachdb/cockroach/commits/8b6f3c84cc256debeeb4d4055c5f0d5c9a481213): ``` === RUN TestStrictGCEnforcement test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestStrictGCEnforcement865385043 test_log_scope.go:74: use -show-logs to present logs inline === CONT TestStrictGCEnforcement client_replica_test.go:3277: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestStrictGCEnforcement865385043 --- FAIL: TestStrictGCEnforcement (0.58s) === RUN TestStrictGCEnforcement/protected_timestamps_are_respected client_replica_test.go:3275: Error Trace: client_replica_test.go:3148 client_replica_test.go:3275 Error: Received unexpected error: batch timestamp 1613803810.475835111,0 must be after replica GC threshold 1613803819.542992787,0 Test: TestStrictGCEnforcement/protected_timestamps_are_respected E210220 06:50:20.549634 490766 kv/kvclient/kvcoord/txn_interceptor_committer.go:449 [-] 1 making txn commit explicit failed for "sql txn" meta={id=e4613806 pri=0.00440269 epo=0 ts=1613803820.543560449,0 min=1613803820.543560449,0 seq=6} lock=true stat=STAGING rts=1613803820.543560449,0 wto=false gul=1613803821.043560449,0 ifw=5: result is ambiguous (server shutdown) --- FAIL: TestStrictGCEnforcement/protected_timestamps_are_respected (0.02s) ``` <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-parallel=4 ``` make stressrace TESTS=TestStrictGCEnforcement PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStrictGCEnforcement.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
kv kvserver teststrictgcenforcement failed on run teststrictgcenforcement test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline cont teststrictgcenforcement client replica test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail teststrictgcenforcement run teststrictgcenforcement protected timestamps are respected client replica test go error trace client replica test go client replica test go error received unexpected error batch timestamp must be after replica gc threshold test teststrictgcenforcement protected timestamps are respected kv kvclient kvcoord txn interceptor committer go making txn commit explicit failed for sql txn meta id pri epo ts min seq lock true stat staging rts wto false gul ifw result is ambiguous server shutdown fail teststrictgcenforcement protected timestamps are respected more parameters tags goflags parallel make stressrace tests teststrictgcenforcement pkg pkg kv kvserver testtimeout stressflags timeout powered by
1
769,091
26,993,203,731
IssuesEvent
2023-02-09 21:48:38
jbx-protocol/juice-interface
https://api.github.com/repos/jbx-protocol/juice-interface
closed
Downloading payments CSV sometimes fails
type:bug priority:2
## Summary Getting this error intermittently: https://user-images.githubusercontent.com/96150256/186579263-e7e045ae-bd65-4c97-9111-348b7e642940.mp4 <img width="1137" alt="Screen Shot 2022-08-25 at 3 02 35 pm" src="https://user-images.githubusercontent.com/96150256/186579240-f095be5b-1162-4e03-bdd9-7671f8f9f468.png"> (I see subgraph I click @peripheralist 😅) ## Steps to reproduce 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ## What is the _current_ bug behavior? ## What is the expected _correct_ behavior? ## Relevant logs and/or screenshots ## Environment - **Device/OS**: [e.g. iOS] - **Browser:** [e.g. chrome, safari] - **Hardware wallet**: [e.g. Ledger w/ Metamask]
1.0
Downloading payments CSV sometimes fails - ## Summary Getting this error intermittently: https://user-images.githubusercontent.com/96150256/186579263-e7e045ae-bd65-4c97-9111-348b7e642940.mp4 <img width="1137" alt="Screen Shot 2022-08-25 at 3 02 35 pm" src="https://user-images.githubusercontent.com/96150256/186579240-f095be5b-1162-4e03-bdd9-7671f8f9f468.png"> (I see subgraph I click @peripheralist 😅) ## Steps to reproduce 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ## What is the _current_ bug behavior? ## What is the expected _correct_ behavior? ## Relevant logs and/or screenshots ## Environment - **Device/OS**: [e.g. iOS] - **Browser:** [e.g. chrome, safari] - **Hardware wallet**: [e.g. Ledger w/ Metamask]
non_test
downloading payments csv sometimes fails summary getting this error intermittently img width alt screen shot at pm src i see subgraph i click peripheralist 😅 steps to reproduce go to click on scroll down to see error what is the current bug behavior what is the expected correct behavior relevant logs and or screenshots environment device os browser hardware wallet
0
102,989
12,836,146,044
IssuesEvent
2020-07-07 13:57:06
trezor/trezor-suite
https://api.github.com/repos/trezor/trezor-suite
closed
Accounts: Jumpy content
Design
It's a nitpick and I kind of understand why it's happening (rollup disappears when selecting Segwit account because we don't allow "non-selected" account). Let's discuss options if there is no obvious solution. (E.g. disable the rollup instead of removing it?) ![Screen Recording 2020-07-02 at 11 10 52](https://user-images.githubusercontent.com/29627086/86340946-e9dc6a80-bc55-11ea-9bd4-6728aa336c4b.gif)
1.0
Accounts: Jumpy content - It's a nitpick and I kind of understand why it's happening (rollup disappears when selecting Segwit account because we don't allow "non-selected" account). Let's discuss options if there is no obvious solution. (E.g. disable the rollup instead of removing it?) ![Screen Recording 2020-07-02 at 11 10 52](https://user-images.githubusercontent.com/29627086/86340946-e9dc6a80-bc55-11ea-9bd4-6728aa336c4b.gif)
non_test
accounts jumpy content it s a nitpick and i kind of understand why it s happening rollup disappears when selecting segwit account because we don t allow non selected account let s discuss options if there is no obvious solution e g disable the rollup instead of removing it
0
416,776
12,151,541,998
IssuesEvent
2020-04-24 20:10:45
jetstack/cert-manager
https://api.github.com/repos/jetstack/cert-manager
closed
Ingress-shim fails to use class of Ingress when creating http01 solver
area/acme/http01 area/ingress-shim kind/bug priority/backlog
**Describe the bug**: Using `cert-manager` v0.11.0. Certificates created by the `ingressShim` in response to the `kubernetes.io/tls-acme: "true"` annotation on an Ingress fail to be issued when the default solver is `http01` and the Ingress has a `kubernetes.io/ingress.class` annotation specifying a class _other than the default_. Instead of using the `ingress.class` of the Ingress requesting the certificate, the Solver uses the default ingress class. **Expected behavior**: According to [the documentation](https://docs.cert-manager.io/en/release-0.11/tasks/issuing-certificates/ingress-shim.html#supported-annotations): > `acme.cert-manager.io/http01-ingress-class` - this annotation allows you to configure ingress class that will be used to solve challenges for this ingress. Customising this is useful when you are trying to secure internal services, and need to solve challenges using different ingress class to that of the ingress. **If not specified and the ‘acme-http01-edit-in-place’ annotation is not set, this defaults to the ingress class of the ingress resource**. _(emphasis added)_ So I expect an Ingress like this: ``` apiVersion: extensions/v1beta1 kind: Ingress metadata: annotations: kubernetes.io/ingress.class: secondary kubernetes.io/tls-acme: "true" spec: rules: - host: cert-test-secondary.example.com http: paths: - backend: serviceName: cert-test-demo servicePort: http path: / tls: - hosts: - cert-test-secondary.example.com secretName: cert-test-tls ``` to generate an `http01` solver using `kubernetes.io/ingress.class: secondary`. Instead (because we have not specified a default `ingress.class` for the default solver), it generates an Ingress solver that has no `kubernetes.io/ingress.class` annotation at all. **Steps to reproduce the bug**: Install `cert-manager` v0.11.0 on Kubernetes 1.12.10 using `helm` chart v0.11.0 with ``` --set ingressShim.defaultIssuerName=letsencrypt-production \ --set ingressShim.defaultIssuerKind=ClusterIssuer ``` See above for Ingress spec. ClusterIssuer: ``` apiVersion: cert-manager.io/v1alpha2 kind: ClusterIssuer metadata: name: letsencrypt-production spec: acme: email: tls@example.com privateKeySecretRef: name: letsencrypt-production-private-key server: https://acme-v02.api.letsencrypt.org/directory solvers: - http01: ingress: {} - dns01: route53: region: us-west-2 selector: matchLabels: cert-manager-solver: dns01 ``` Note that if we specify an Ingress class in the Solver, like ``` - http01: ingress: class: none ``` Then the challenge solver ingress gets created with that `ingress.class`, not the one of the target ingress. **Anything else we need to know?**: We do not have a default ingress installed, nor do we have an easy mapping of domain names to ingresses to allow the ClusterIssuer to pick the right ingress based on the domain names of the certificate. We expect the Challenge Solver to use the Ingress class of the Ingress where it will be deployed, as described in the documentation. The workaround is to add the `acme.cert-manager.io/http01-ingress-class` annotation to the Ingress, but that requires us to update all of our Ingresses and all of the `helm` charts/values that install them, which is highly undesirable. **Environment details:**: - Kubernetes version: v1.12.10 - Cloud-provider/provisioner: AWS - cert-manager version : v0.11.0 - Install method: helm /kind bug
1.0
Ingress-shim fails to use class of Ingress when creating http01 solver - **Describe the bug**: Using `cert-manager` v0.11.0. Certificates created by the `ingressShim` in response to the `kubernetes.io/tls-acme: "true"` annotation on an Ingress fail to be issued when the default solver is `http01` and the Ingress has a `kubernetes.io/ingress.class` annotation specifying a class _other than the default_. Instead of using the `ingress.class` of the Ingress requesting the certificate, the Solver uses the default ingress class. **Expected behavior**: According to [the documentation](https://docs.cert-manager.io/en/release-0.11/tasks/issuing-certificates/ingress-shim.html#supported-annotations): > `acme.cert-manager.io/http01-ingress-class` - this annotation allows you to configure ingress class that will be used to solve challenges for this ingress. Customising this is useful when you are trying to secure internal services, and need to solve challenges using different ingress class to that of the ingress. **If not specified and the ‘acme-http01-edit-in-place’ annotation is not set, this defaults to the ingress class of the ingress resource**. _(emphasis added)_ So I expect an Ingress like this: ``` apiVersion: extensions/v1beta1 kind: Ingress metadata: annotations: kubernetes.io/ingress.class: secondary kubernetes.io/tls-acme: "true" spec: rules: - host: cert-test-secondary.example.com http: paths: - backend: serviceName: cert-test-demo servicePort: http path: / tls: - hosts: - cert-test-secondary.example.com secretName: cert-test-tls ``` to generate an `http01` solver using `kubernetes.io/ingress.class: secondary`. Instead (because we have not specified a default `ingress.class` for the default solver), it generates an Ingress solver that has no `kubernetes.io/ingress.class` annotation at all. **Steps to reproduce the bug**: Install `cert-manager` v0.11.0 on Kubernetes 1.12.10 using `helm` chart v0.11.0 with ``` --set ingressShim.defaultIssuerName=letsencrypt-production \ --set ingressShim.defaultIssuerKind=ClusterIssuer ``` See above for Ingress spec. ClusterIssuer: ``` apiVersion: cert-manager.io/v1alpha2 kind: ClusterIssuer metadata: name: letsencrypt-production spec: acme: email: tls@example.com privateKeySecretRef: name: letsencrypt-production-private-key server: https://acme-v02.api.letsencrypt.org/directory solvers: - http01: ingress: {} - dns01: route53: region: us-west-2 selector: matchLabels: cert-manager-solver: dns01 ``` Note that if we specify an Ingress class in the Solver, like ``` - http01: ingress: class: none ``` Then the challenge solver ingress gets created with that `ingress.class`, not the one of the target ingress. **Anything else we need to know?**: We do not have a default ingress installed, nor do we have an easy mapping of domain names to ingresses to allow the ClusterIssuer to pick the right ingress based on the domain names of the certificate. We expect the Challenge Solver to use the Ingress class of the Ingress where it will be deployed, as described in the documentation. The workaround is to add the `acme.cert-manager.io/http01-ingress-class` annotation to the Ingress, but that requires us to update all of our Ingresses and all of the `helm` charts/values that install them, which is highly undesirable. **Environment details:**: - Kubernetes version: v1.12.10 - Cloud-provider/provisioner: AWS - cert-manager version : v0.11.0 - Install method: helm /kind bug
non_test
ingress shim fails to use class of ingress when creating solver describe the bug using cert manager certificates created by the ingressshim in response to the kubernetes io tls acme true annotation on an ingress fail to be issued when the default solver is and the ingress has a kubernetes io ingress class annotation specifying a class other than the default instead of using the ingress class of the ingress requesting the certificate the solver uses the default ingress class expected behavior according to acme cert manager io ingress class this annotation allows you to configure ingress class that will be used to solve challenges for this ingress customising this is useful when you are trying to secure internal services and need to solve challenges using different ingress class to that of the ingress if not specified and the ‘acme edit in place’ annotation is not set this defaults to the ingress class of the ingress resource emphasis added so i expect an ingress like this apiversion extensions kind ingress metadata annotations kubernetes io ingress class secondary kubernetes io tls acme true spec rules host cert test secondary example com http paths backend servicename cert test demo serviceport http path tls hosts cert test secondary example com secretname cert test tls to generate an solver using kubernetes io ingress class secondary instead because we have not specified a default ingress class for the default solver it generates an ingress solver that has no kubernetes io ingress class annotation at all steps to reproduce the bug install cert manager on kubernetes using helm chart with set ingressshim defaultissuername letsencrypt production set ingressshim defaultissuerkind clusterissuer see above for ingress spec clusterissuer apiversion cert manager io kind clusterissuer metadata name letsencrypt production spec acme email tls example com privatekeysecretref name letsencrypt production private key server solvers ingress region us west selector matchlabels cert manager solver note that if we specify an ingress class in the solver like ingress class none then the challenge solver ingress gets created with that ingress class not the one of the target ingress anything else we need to know we do not have a default ingress installed nor do we have an easy mapping of domain names to ingresses to allow the clusterissuer to pick the right ingress based on the domain names of the certificate we expect the challenge solver to use the ingress class of the ingress where it will be deployed as described in the documentation the workaround is to add the acme cert manager io ingress class annotation to the ingress but that requires us to update all of our ingresses and all of the helm charts values that install them which is highly undesirable environment details kubernetes version cloud provider provisioner aws cert manager version install method helm kind bug
0
82,283
7,836,299,014
IssuesEvent
2018-06-17 17:44:52
vmware/vic
https://api.github.com/repos/vmware/vic
opened
Infra: parallelism: TLS certificates deleted mid test
area/infra component/test
This was found in the following CI build with parallelism of 6. CI build: https://ci-vic.vmware.com/vmware/vic/19383/7 ``` KEYWORD ${rc}, ${output} = OperatingSystem . Run And Return Rc And Output docker %{VCH-PARAMS} start ${container} ${rc} = 0 ${output} = 6c79317fccbd061f6669c5d1abb6fc1919cbaf0ecaa36c56c74ff7e3af250b24 <snip logging step> KEYWORD ${rc}, ${output} = OperatingSystem . Run And Return Rc And Output docker %{VCH-PARAMS} ps ${rc} = 1 ${output} = could not read CA certificate "/go/src/github.com/vmware/vic/foo-bar-certs/ca.pem": open /go/src/github.com/vmware/vic/foo-bar-certs/ca.pem: no such file or directory ``` The following tests use `foo-bar-certs`: * `6-13-TLS.Create VCH - defaults custom cert path` - test uses that path * `6-17-Configure-TLS` - suite setup uses the path and tests explicitly delete it.
1.0
Infra: parallelism: TLS certificates deleted mid test - This was found in the following CI build with parallelism of 6. CI build: https://ci-vic.vmware.com/vmware/vic/19383/7 ``` KEYWORD ${rc}, ${output} = OperatingSystem . Run And Return Rc And Output docker %{VCH-PARAMS} start ${container} ${rc} = 0 ${output} = 6c79317fccbd061f6669c5d1abb6fc1919cbaf0ecaa36c56c74ff7e3af250b24 <snip logging step> KEYWORD ${rc}, ${output} = OperatingSystem . Run And Return Rc And Output docker %{VCH-PARAMS} ps ${rc} = 1 ${output} = could not read CA certificate "/go/src/github.com/vmware/vic/foo-bar-certs/ca.pem": open /go/src/github.com/vmware/vic/foo-bar-certs/ca.pem: no such file or directory ``` The following tests use `foo-bar-certs`: * `6-13-TLS.Create VCH - defaults custom cert path` - test uses that path * `6-17-Configure-TLS` - suite setup uses the path and tests explicitly delete it.
test
infra parallelism tls certificates deleted mid test this was found in the following ci build with parallelism of ci build keyword rc output operatingsystem run and return rc and output docker vch params start container rc output keyword rc output operatingsystem run and return rc and output docker vch params ps rc output could not read ca certificate go src github com vmware vic foo bar certs ca pem open go src github com vmware vic foo bar certs ca pem no such file or directory the following tests use foo bar certs tls create vch defaults custom cert path test uses that path configure tls suite setup uses the path and tests explicitly delete it
1
245,307
20,760,499,728
IssuesEvent
2022-03-15 15:46:10
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
closed
Investigate and fix failure of selected buttons from automated tests
testing task e2e-tests
Discovered failures in steps checking if a button from a group of options was selected. Example: https://jenkins.sormas.netzlink.com/view/SORMAS-Project/job/sormas-RunE2eTests/562/allure/#suites/ba64af4344a5d1b16948f67eae90e934/8a35195df3e9b1f2/ Checked and the locator is correct in debug/manual check, the problem is only from automated test execution. Investigation starting point: `private String getTaskContext() { return webDriverHelpers.getCheckedOptionFromHorizontalOptionGroup(SELECTED_TASK_CONTEXT); }` Suggestions: Try to update selenium version Try to update browse's version Try to discuss with UI developers if a checked tag can be added to component Refactor all code to get all options from a group of buttons and look after selected one
2.0
Investigate and fix failure of selected buttons from automated tests - Discovered failures in steps checking if a button from a group of options was selected. Example: https://jenkins.sormas.netzlink.com/view/SORMAS-Project/job/sormas-RunE2eTests/562/allure/#suites/ba64af4344a5d1b16948f67eae90e934/8a35195df3e9b1f2/ Checked and the locator is correct in debug/manual check, the problem is only from automated test execution. Investigation starting point: `private String getTaskContext() { return webDriverHelpers.getCheckedOptionFromHorizontalOptionGroup(SELECTED_TASK_CONTEXT); }` Suggestions: Try to update selenium version Try to update browse's version Try to discuss with UI developers if a checked tag can be added to component Refactor all code to get all options from a group of buttons and look after selected one
test
investigate and fix failure of selected buttons from automated tests discovered failures in steps checking if a button from a group of options was selected example checked and the locator is correct in debug manual check the problem is only from automated test execution investigation starting point private string gettaskcontext return webdriverhelpers getcheckedoptionfromhorizontaloptiongroup selected task context suggestions try to update selenium version try to update browse s version try to discuss with ui developers if a checked tag can be added to component refactor all code to get all options from a group of buttons and look after selected one
1
107,902
9,247,884,773
IssuesEvent
2019-03-15 02:59:57
open-apparel-registry/open-apparel-registry
https://api.github.com/repos/open-apparel-registry/open-apparel-registry
closed
Update fixture generator and committed fixtures to be more realistic
tested/verified
## Overview We will soon be taking screen captures for promotional videos and asking for design feedback from some trusted contributors. Attempt to avoid questions and confusion by making the fixture data more closely resemble actual data. ### Describe the solution you'd like - Geocode the facility address so that they show up at real locations rather than randomized points around Philladelpha. - Generate contributor names that match the type of the contributor, rather than random names i.e. Brand A, Brand B, Manufacturer Group A, Service Provider B - Create a roughly equal distribution of contributors in each type.
1.0
Update fixture generator and committed fixtures to be more realistic - ## Overview We will soon be taking screen captures for promotional videos and asking for design feedback from some trusted contributors. Attempt to avoid questions and confusion by making the fixture data more closely resemble actual data. ### Describe the solution you'd like - Geocode the facility address so that they show up at real locations rather than randomized points around Philladelpha. - Generate contributor names that match the type of the contributor, rather than random names i.e. Brand A, Brand B, Manufacturer Group A, Service Provider B - Create a roughly equal distribution of contributors in each type.
test
update fixture generator and committed fixtures to be more realistic overview we will soon be taking screen captures for promotional videos and asking for design feedback from some trusted contributors attempt to avoid questions and confusion by making the fixture data more closely resemble actual data describe the solution you d like geocode the facility address so that they show up at real locations rather than randomized points around philladelpha generate contributor names that match the type of the contributor rather than random names i e brand a brand b manufacturer group a service provider b create a roughly equal distribution of contributors in each type
1
96,092
8,586,089,280
IssuesEvent
2018-11-14 06:51:46
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Failing test: [sig-autoscaling] [HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] ReplicationController Should scale from 1 pod to 3 pods and from 3 to 5 and verify decision stabilityChanges
kind/failing-test sig/autoscaling
<!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs --> **Which jobs are failing**: [sig-release-master-upgrade/gce-new-master-upgrade-master](https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-new-master-upgrade-master) **Which test(s) are failing**: [[HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] ReplicationController Should scale from 1 pod to 3 pods and from 3 to 5 and verify decision stability](https://gubernator.k8s.io/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-master/1862#sig-autoscaling-hpa-horizontal-pod-autoscaling-scale-resource-cpu-sig-autoscaling-serial-slow-replicationcontroller-should-scale-from-1-pod-to-3-pods-and-from-3-to-5-and-verify-decision-stability) **Since when has it been failing**: November 3rd, 2018 **Testgrid link**: https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-new-master-upgrade-master https://gubernator.k8s.io/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-master/1862#sig-autoscaling-hpa-horizontal-pod-autoscaling-scale-resource-cpu-sig-autoscaling-serial-slow-replicationcontroller-should-scale-from-1-pod-to-3-pods-and-from-3-to-5-and-verify-decision-stability **Anything else we need to know**: [diff towards last passing commit](https://github.com/kubernetes/kubernetes/compare/3293f02d0...46d173232?) hints at #69698 . <!-- DO NOT EDIT BELOW THIS LINE --> /kind failing-test
1.0
Failing test: [sig-autoscaling] [HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] ReplicationController Should scale from 1 pod to 3 pods and from 3 to 5 and verify decision stabilityChanges - <!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs --> **Which jobs are failing**: [sig-release-master-upgrade/gce-new-master-upgrade-master](https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-new-master-upgrade-master) **Which test(s) are failing**: [[HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] ReplicationController Should scale from 1 pod to 3 pods and from 3 to 5 and verify decision stability](https://gubernator.k8s.io/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-master/1862#sig-autoscaling-hpa-horizontal-pod-autoscaling-scale-resource-cpu-sig-autoscaling-serial-slow-replicationcontroller-should-scale-from-1-pod-to-3-pods-and-from-3-to-5-and-verify-decision-stability) **Since when has it been failing**: November 3rd, 2018 **Testgrid link**: https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-new-master-upgrade-master https://gubernator.k8s.io/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-master/1862#sig-autoscaling-hpa-horizontal-pod-autoscaling-scale-resource-cpu-sig-autoscaling-serial-slow-replicationcontroller-should-scale-from-1-pod-to-3-pods-and-from-3-to-5-and-verify-decision-stability **Anything else we need to know**: [diff towards last passing commit](https://github.com/kubernetes/kubernetes/compare/3293f02d0...46d173232?) hints at #69698 . <!-- DO NOT EDIT BELOW THIS LINE --> /kind failing-test
test
failing test horizontal pod autoscaling scale resource cpu replicationcontroller should scale from pod to pods and from to and verify decision stabilitychanges which jobs are failing which test s are failing horizontal pod autoscaling scale resource cpu replicationcontroller should scale from pod to pods and from to and verify decision stability since when has it been failing november testgrid link anything else we need to know hints at kind failing test
1
139,319
5,366,935,598
IssuesEvent
2017-02-22 01:55:16
chrisblakley/Nebula
https://api.github.com/repos/chrisblakley/Nebula
closed
Consider better organizing of theme file structure.
Backend (Server) Feature Request Frontend (Script) Frontend (Style) Low Priority Parent / Child Theme
I see a lot of pro themes has their client side resources located in assets directory and each one in their respectives folders and one called lib to store all external resources We could keep this folder structure too (also on nebula integrations) Of course we need to update sass compiler too
1.0
Consider better organizing of theme file structure. - I see a lot of pro themes has their client side resources located in assets directory and each one in their respectives folders and one called lib to store all external resources We could keep this folder structure too (also on nebula integrations) Of course we need to update sass compiler too
non_test
consider better organizing of theme file structure i see a lot of pro themes has their client side resources located in assets directory and each one in their respectives folders and one called lib to store all external resources we could keep this folder structure too also on nebula integrations of course we need to update sass compiler too
0
40,597
8,815,302,383
IssuesEvent
2018-12-29 16:51:11
JWWeatherman/math_bot
https://api.github.com/repos/JWWeatherman/math_bot
opened
Implement Sounds
front end code
### Robot animation sounds - [x] Wall Bump Sound - [x] Walk Sound - [ ] Turn Sound - [ ] Pickup Sound - [ ] Drop Sound ### Star-system background sounds - [x] Same Sound For All - [ ] Get Different Sound For Each
1.0
Implement Sounds - ### Robot animation sounds - [x] Wall Bump Sound - [x] Walk Sound - [ ] Turn Sound - [ ] Pickup Sound - [ ] Drop Sound ### Star-system background sounds - [x] Same Sound For All - [ ] Get Different Sound For Each
non_test
implement sounds robot animation sounds wall bump sound walk sound turn sound pickup sound drop sound star system background sounds same sound for all get different sound for each
0
14,556
9,336,607,154
IssuesEvent
2019-03-28 21:44:27
codevirtuel/osu-replayViewer-web
https://api.github.com/repos/codevirtuel/osu-replayViewer-web
closed
CVE-2018-19838 Medium Severity Vulnerability detected by WhiteSource
security vulnerability
## CVE-2018-19838 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> </p> </details> </p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/expand.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/color_maps.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_util.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/output.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_values.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/util.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/emitter.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/lexer.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_node.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/plugins.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/base.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/position.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/subset_map.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operation.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/error_handling.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_importer_bridge.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/functions.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/eval.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_context_wrapper.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/error_handling.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/node.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/parser.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/subset_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/emitter.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/listize.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/output.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/functions.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cssize.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/prelexer.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/paths.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/inspect.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/color.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_unification.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/values.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_util.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/source_map.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/list.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/json.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/units.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/units.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/context.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8/checked.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/listize.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/string.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/prelexer.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/context.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/boolean.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass2scss.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/eval.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/expand.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/factory.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operators.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/boolean.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/source_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/value.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/callback_bridge.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/file.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/node.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/environment.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/extend.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_context.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operators.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/constants.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/parser.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/constants.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/list.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cssize.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/functions.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/util.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_function_bridge.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_importer_bridge.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/bind.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/inspect.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/backtrace.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/extend.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/debugger.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cencode.c - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/number.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/color.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/c99func.c - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/position.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_values.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/values.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/null.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/context.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_c.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_value.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/color_maps.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_context_wrapper.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/script/test-leaks.pl - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/lexer.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_c.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_value.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/b64/encode.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/file.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/environment.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/plugins.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/binding.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_context.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass prior to 3.5.5, functions inside ast.cpp for IMPLEMENT_AST_OPERATORS expansion allow attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, as demonstrated by recursive calls involving clone(), cloneChildren(), and copy(). <p>Publish Date: 2018-12-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838>CVE-2018-19838</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838</a></p> <p>Fix Resolution: 3.5.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19838 Medium Severity Vulnerability detected by WhiteSource - ## CVE-2018-19838 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> </p> </details> </p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/expand.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/color_maps.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_util.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/output.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_values.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/util.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/emitter.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/lexer.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_node.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/plugins.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/base.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/position.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/subset_map.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operation.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/error_handling.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_importer_bridge.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/functions.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/eval.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_context_wrapper.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/error_handling.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/node.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/parser.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/subset_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/emitter.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/listize.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/output.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/functions.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cssize.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/prelexer.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/paths.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/inspect.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/color.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_unification.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/values.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_util.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/source_map.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/list.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/json.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/units.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/units.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/context.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8/checked.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/listize.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/string.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/prelexer.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/context.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/boolean.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass2scss.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/eval.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/expand.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/factory.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operators.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/boolean.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/source_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/value.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/callback_bridge.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/file.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/node.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/environment.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/extend.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_context.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/operators.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/constants.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/parser.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/constants.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/list.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cssize.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/functions.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/util.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_function_bridge.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/custom_importer_bridge.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/bind.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/inspect.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/backtrace.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/extend.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/debugger.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/cencode.c - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/number.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/color.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/c99func.c - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/position.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_values.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/values.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/null.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/ast.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/include/sass/context.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_c.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_value.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/color_maps.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_context_wrapper.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/script/test-leaks.pl - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/lexer.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_c.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/sass_types/map.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/to_value.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/b64/encode.h - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/file.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/environment.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/plugins.hpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/binding.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/sass_context.cpp - /osu-replayViewer-web/bulma/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass prior to 3.5.5, functions inside ast.cpp for IMPLEMENT_AST_OPERATORS expansion allow attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, as demonstrated by recursive calls involving clone(), cloneChildren(), and copy(). <p>Publish Date: 2018-12-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838>CVE-2018-19838</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19838</a></p> <p>Fix Resolution: 3.5.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium severity vulnerability detected by whitesource cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries osu replayviewer web bulma node modules node sass src libsass src expand hpp osu replayviewer web bulma node modules node sass src libsass src color maps cpp osu replayviewer web bulma node modules node sass src libsass src sass util hpp osu replayviewer web bulma node modules node sass src libsass src unchecked h osu replayviewer web bulma node modules node sass src libsass src output hpp osu replayviewer web bulma node modules node sass src libsass src sass values hpp osu replayviewer web bulma node modules node sass src libsass src util hpp osu replayviewer web bulma node modules node sass src libsass src emitter hpp osu replayviewer web bulma node modules node sass src libsass src lexer cpp osu replayviewer web bulma node modules node sass src libsass test test node cpp osu replayviewer web bulma node modules node sass src libsass src plugins cpp osu replayviewer web bulma node modules node sass src libsass include sass base h osu replayviewer web bulma node modules node sass src libsass src position hpp osu replayviewer web bulma node modules node sass src libsass src subset map hpp osu replayviewer web bulma node modules node sass src libsass src operation hpp osu replayviewer web bulma node modules node sass src libsass src remove placeholders cpp osu replayviewer web bulma node modules node sass src libsass src error handling hpp osu replayviewer web bulma node modules node sass src custom importer bridge cpp osu replayviewer web bulma node modules node sass src libsass contrib plugin cpp osu replayviewer web bulma node modules node sass src libsass src functions hpp osu replayviewer web bulma node modules node sass src libsass test test superselector cpp osu replayviewer web bulma node modules node sass src libsass src eval hpp osu replayviewer web bulma node modules node sass src libsass src string hpp osu replayviewer web bulma node modules node sass src sass context wrapper h osu replayviewer web bulma node modules node sass src libsass src error handling cpp osu replayviewer web bulma node modules node sass src libsass src node cpp osu replayviewer web bulma node modules node sass src libsass src parser cpp osu replayviewer web bulma node modules node sass src libsass src subset map cpp osu replayviewer web bulma node modules node sass src libsass src emitter cpp osu replayviewer web bulma node modules node sass src libsass src listize cpp osu replayviewer web bulma node modules node sass src libsass src ast hpp osu replayviewer web bulma node modules node sass src libsass src sass functions hpp osu replayviewer web bulma node modules node sass src libsass src memory sharedptr cpp osu replayviewer web bulma node modules node sass src libsass src output cpp osu replayviewer web bulma node modules node sass src libsass src check nesting cpp osu replayviewer web bulma node modules node sass src libsass src ast def macros hpp osu replayviewer web bulma node modules node sass src libsass src functions cpp osu replayviewer web bulma node modules node sass src libsass src cssize hpp osu replayviewer web bulma node modules node sass src libsass src prelexer cpp osu replayviewer web bulma node modules node sass src libsass src paths hpp osu replayviewer web bulma node modules node sass src libsass src ast fwd decl hpp osu replayviewer web bulma node modules node sass src libsass src inspect hpp osu replayviewer web bulma node modules node sass src sass types color cpp osu replayviewer web bulma node modules node sass src libsass test test unification cpp osu replayviewer web bulma node modules node sass src libsass src values cpp osu replayviewer web bulma node modules node sass src libsass src sass util cpp osu replayviewer web bulma node modules node sass src libsass src source map hpp osu replayviewer web bulma node modules node sass src sass types list h osu replayviewer web bulma node modules node sass src libsass src check nesting hpp osu replayviewer web bulma node modules node sass src libsass src json cpp osu replayviewer web bulma node modules node sass src libsass src units cpp osu replayviewer web bulma node modules node sass src libsass src units hpp osu replayviewer web bulma node modules node sass src libsass src context cpp osu replayviewer web bulma node modules node sass src libsass src checked h osu replayviewer web bulma node modules node sass src libsass src listize hpp osu replayviewer web bulma node modules node sass src sass types string cpp osu replayviewer web bulma node modules node sass src libsass src prelexer hpp osu replayviewer web bulma node modules node sass src libsass src context hpp osu replayviewer web bulma node modules node sass src sass types boolean h osu replayviewer web bulma node modules node sass src libsass include h osu replayviewer web bulma node modules node sass src libsass src eval cpp osu replayviewer web bulma node modules node sass src libsass src expand cpp osu replayviewer web bulma node modules node sass src sass types factory cpp osu replayviewer web bulma node modules node sass src libsass src operators cpp osu replayviewer web bulma node modules node sass src sass types boolean cpp osu replayviewer web bulma node modules node sass src libsass src source map cpp osu replayviewer web bulma node modules node sass src sass types value h osu replayviewer web bulma node modules node sass src libsass src string cpp osu replayviewer web bulma node modules node sass src callback bridge h osu replayviewer web bulma node modules node sass src libsass src file cpp osu replayviewer web bulma node modules node sass src libsass src sass cpp osu replayviewer web bulma node modules node sass src libsass src node hpp osu replayviewer web bulma node modules node sass src libsass src environment cpp osu replayviewer web bulma node modules node sass src libsass src extend hpp osu replayviewer web bulma node modules node sass src libsass src sass context hpp osu replayviewer web bulma node modules node sass src libsass src operators hpp osu replayviewer web bulma node modules node sass src libsass src constants hpp osu replayviewer web bulma node modules node sass src libsass src sass hpp osu replayviewer web bulma node modules node sass src libsass src ast fwd decl cpp osu replayviewer web bulma node modules node sass src libsass src parser hpp osu replayviewer web bulma node modules node sass src libsass src constants cpp osu replayviewer web bulma node modules node sass src sass types list cpp osu replayviewer web bulma node modules node sass src libsass src cssize cpp osu replayviewer web bulma node modules node sass src libsass include sass functions h osu replayviewer web bulma node modules node sass src libsass src util cpp osu replayviewer web bulma node modules node sass src custom function bridge cpp osu replayviewer web bulma node modules node sass src custom importer bridge h osu replayviewer web bulma node modules node sass src libsass src bind cpp osu replayviewer web bulma node modules node sass src libsass src inspect cpp osu replayviewer web bulma node modules node sass src libsass src sass functions cpp osu replayviewer web bulma node modules node sass src libsass src backtrace cpp osu replayviewer web bulma node modules node sass src libsass src extend cpp osu replayviewer web bulma node modules node sass src sass types sass value wrapper h osu replayviewer web bulma node modules node sass src libsass src debugger hpp osu replayviewer web bulma node modules node sass src libsass src cencode c osu replayviewer web bulma node modules node sass src libsass src cpp osu replayviewer web bulma node modules node sass src sass types number cpp osu replayviewer web bulma node modules node sass src sass types color h osu replayviewer web bulma node modules node sass src libsass src c osu replayviewer web bulma node modules node sass src libsass src position cpp osu replayviewer web bulma node modules node sass src libsass src remove placeholders hpp osu replayviewer web bulma node modules node sass src libsass src sass values cpp osu replayviewer web bulma node modules node sass src libsass include sass values h osu replayviewer web bulma node modules node sass src libsass test test subset map cpp osu replayviewer web bulma node modules node sass src libsass src cpp osu replayviewer web bulma node modules node sass src sass types null cpp osu replayviewer web bulma node modules node sass src libsass src ast cpp osu replayviewer web bulma node modules node sass src libsass include sass context h osu replayviewer web bulma node modules node sass src libsass src to c cpp osu replayviewer web bulma node modules node sass src libsass src to value hpp osu replayviewer web bulma node modules node sass src libsass src color maps hpp osu replayviewer web bulma node modules node sass src sass context wrapper cpp osu replayviewer web bulma node modules node sass src libsass script test leaks pl osu replayviewer web bulma node modules node sass src libsass src lexer hpp osu replayviewer web bulma node modules node sass src libsass src memory sharedptr hpp osu replayviewer web bulma node modules node sass src libsass src to c hpp osu replayviewer web bulma node modules node sass src sass types map cpp osu replayviewer web bulma node modules node sass src libsass src to value cpp osu replayviewer web bulma node modules node sass src libsass src encode h osu replayviewer web bulma node modules node sass src libsass src file hpp osu replayviewer web bulma node modules node sass src libsass src environment hpp osu replayviewer web bulma node modules node sass src libsass src plugins hpp osu replayviewer web bulma node modules node sass src binding cpp osu replayviewer web bulma node modules node sass src libsass src sass context cpp osu replayviewer web bulma node modules node sass src libsass src debug hpp vulnerability details in libsass prior to functions inside ast cpp for implement ast operators expansion allow attackers to cause a denial of service resulting from stack consumption via a crafted sass file as demonstrated by recursive calls involving clone clonechildren and copy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution step up your open source security game with whitesource
0
101,455
16,511,654,137
IssuesEvent
2021-05-26 05:25:17
kijunb33/test
https://api.github.com/repos/kijunb33/test
opened
CVE-2021-25329 (High) detected in tomcat-embed-core-7.0.90.jar
security vulnerability
## CVE-2021-25329 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-7.0.90.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to vulnerable library: test/tomcat-embed-core-7.0.90.jar</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-7.0.90.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/kijunb33/test/commits/8df5c209ab0589b3f881b1e4a6c004c81ae3d659">8df5c209ab0589b3f881b1e4a6c004c81ae3d659</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The fix for CVE-2020-9484 was incomplete. When using Apache Tomcat 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41, 8.5.0 to 8.5.61 or 7.0.0. to 7.0.107 with a configuration edge case that was highly unlikely to be used, the Tomcat instance was still vulnerable to CVE-2020-9494. Note that both the previously published prerequisites for CVE-2020-9484 and the previously published mitigations for CVE-2020-9484 also apply to this issue. <p>Publish Date: 2021-03-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25329>CVE-2021-25329</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/rfe62fbf9d4c314f166fe8c668e50e5d9dd882a99447f26f0367474bf%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rfe62fbf9d4c314f166fe8c668e50e5d9dd882a99447f26f0367474bf%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2021-03-01</p> <p>Fix Resolution: org.apache.tomcat:tomcat:7.0.108, org.apache.tomcat:tomcat:8.5.63, org.apache.tomcat:tomcat:9.0.43,org.apache.tomcat:tomcat:10.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-25329 (High) detected in tomcat-embed-core-7.0.90.jar - ## CVE-2021-25329 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-7.0.90.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to vulnerable library: test/tomcat-embed-core-7.0.90.jar</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-7.0.90.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/kijunb33/test/commits/8df5c209ab0589b3f881b1e4a6c004c81ae3d659">8df5c209ab0589b3f881b1e4a6c004c81ae3d659</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The fix for CVE-2020-9484 was incomplete. When using Apache Tomcat 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41, 8.5.0 to 8.5.61 or 7.0.0. to 7.0.107 with a configuration edge case that was highly unlikely to be used, the Tomcat instance was still vulnerable to CVE-2020-9494. Note that both the previously published prerequisites for CVE-2020-9484 and the previously published mitigations for CVE-2020-9484 also apply to this issue. <p>Publish Date: 2021-03-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25329>CVE-2021-25329</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/rfe62fbf9d4c314f166fe8c668e50e5d9dd882a99447f26f0367474bf%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rfe62fbf9d4c314f166fe8c668e50e5d9dd882a99447f26f0367474bf%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2021-03-01</p> <p>Fix Resolution: org.apache.tomcat:tomcat:7.0.108, org.apache.tomcat:tomcat:8.5.63, org.apache.tomcat:tomcat:9.0.43,org.apache.tomcat:tomcat:10.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to vulnerable library test tomcat embed core jar dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href found in base branch master vulnerability details the fix for cve was incomplete when using apache tomcat to to to or to with a configuration edge case that was highly unlikely to be used the tomcat instance was still vulnerable to cve note that both the previously published prerequisites for cve and the previously published mitigations for cve also apply to this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat org apache tomcat tomcat org apache tomcat tomcat org apache tomcat tomcat step up your open source security game with whitesource
0
71,641
3,366,122,828
IssuesEvent
2015-11-21 03:21:26
hackthesystemATX/doctrinr
https://api.github.com/repos/hackthesystemATX/doctrinr
reopened
User Flow Diagram
high priority
Someone needs to creat a low-fidelity diagram of the userflow. [Example] (https://www.google.com/search?q=userflow&source=lnms&tbm=isch&sa=X&ved=0ahUKEwjg9KfPxaDJAhWCUogKHTYzC0cQ_AUICCgC&biw=1036&bih=805) Something simple so we can easily communicate it to a desiger/developer.
1.0
User Flow Diagram - Someone needs to creat a low-fidelity diagram of the userflow. [Example] (https://www.google.com/search?q=userflow&source=lnms&tbm=isch&sa=X&ved=0ahUKEwjg9KfPxaDJAhWCUogKHTYzC0cQ_AUICCgC&biw=1036&bih=805) Something simple so we can easily communicate it to a desiger/developer.
non_test
user flow diagram someone needs to creat a low fidelity diagram of the userflow something simple so we can easily communicate it to a desiger developer
0
1,320
3,593,205,104
IssuesEvent
2016-02-01 18:50:36
Azure/azure-powershell
https://api.github.com/repos/Azure/azure-powershell
closed
AzureRmBackupContainer V1VM/V2VM confusion
RecoveryServices
According to the documentation, and the source, there is a Parameter set for v2VMs However when you run the command - Register-AzureRmBackupContainer -Vault $vault -Name $vm.Name ` -ResourceGroupName $vm.ResourceGroupName The error message `The specified Azure Virtual Machine Not Found.` suggests that a possible cause is `VM is Azure IaaS v2 VM` Obviously this goes against what the command line options suggest. Is it possible to backup an IaaS v2 VM with AzureRM.Backup? (considering that it is a RM cmdlet, it would hint towards it being possible!) <!---@tfsbridge:{"tfsId":5629458}-->
1.0
AzureRmBackupContainer V1VM/V2VM confusion - According to the documentation, and the source, there is a Parameter set for v2VMs However when you run the command - Register-AzureRmBackupContainer -Vault $vault -Name $vm.Name ` -ResourceGroupName $vm.ResourceGroupName The error message `The specified Azure Virtual Machine Not Found.` suggests that a possible cause is `VM is Azure IaaS v2 VM` Obviously this goes against what the command line options suggest. Is it possible to backup an IaaS v2 VM with AzureRM.Backup? (considering that it is a RM cmdlet, it would hint towards it being possible!) <!---@tfsbridge:{"tfsId":5629458}-->
non_test
azurermbackupcontainer confusion according to the documentation and the source there is a parameter set for however when you run the command register azurermbackupcontainer vault vault name vm name resourcegroupname vm resourcegroupname the error message the specified azure virtual machine not found suggests that a possible cause is vm is azure iaas vm obviously this goes against what the command line options suggest is it possible to backup an iaas vm with azurerm backup considering that it is a rm cmdlet it would hint towards it being possible
0
154,739
12,226,983,221
IssuesEvent
2020-05-03 13:25:05
RomanKondratev90/Test-project
https://api.github.com/repos/RomanKondratev90/Test-project
opened
Тест - кейс №6 (ручка). Проверка ручки на внешние механические повреждения
test case (Pass)
Приоритет - средний Шаги 1 Взять ручку руку. 2 Провести осмотр ручки Ожидаемы результат Длина - отсутствие повреждений ручки. Статус Положительный результат
1.0
Тест - кейс №6 (ручка). Проверка ручки на внешние механические повреждения - Приоритет - средний Шаги 1 Взять ручку руку. 2 Провести осмотр ручки Ожидаемы результат Длина - отсутствие повреждений ручки. Статус Положительный результат
test
тест кейс № ручка проверка ручки на внешние механические повреждения приоритет средний шаги взять ручку руку провести осмотр ручки ожидаемы результат длина отсутствие повреждений ручки статус положительный результат
1
45,051
5,682,977,852
IssuesEvent
2017-04-13 11:19:03
carbontwelve/tapestry
https://api.github.com/repos/carbontwelve/tapestry
opened
Check that two files of the same permalink don't conflict
bug needs-tests
If you have two files with the same permalink, Tapestry should highlight to the user the error and fail-back to one of two configured modes: 1. Rename the file with an increment number e.g. `filename-1.html` 2. Throw an error and output to console the issue
1.0
Check that two files of the same permalink don't conflict - If you have two files with the same permalink, Tapestry should highlight to the user the error and fail-back to one of two configured modes: 1. Rename the file with an increment number e.g. `filename-1.html` 2. Throw an error and output to console the issue
test
check that two files of the same permalink don t conflict if you have two files with the same permalink tapestry should highlight to the user the error and fail back to one of two configured modes rename the file with an increment number e g filename html throw an error and output to console the issue
1
298,147
25,793,338,747
IssuesEvent
2022-12-10 09:29:28
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
opened
test_unidirectional_bucket_replication failed with ubprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 second
TestCase failing Squad/Red
Run details: URL: https://reportportal-ocs4.apps.ocp-c1.prod.psi.redhat.com/ui/#OCS/launches/362/7364/320363/320416/320422/log Run ID: 1670509372 Test Case: test_unidirectional_bucket_replication ODF Build: 4.12.0-130 OCP Version: 4.12 Job name: IBM Cloud IPI 3AZ RHCOS 3M 3W tier1 Jenkins job: https://ocs4-jenkins-csb-odf-qe.apps.ocp-c1.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/6343/ Logs URL: http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j-005ici3c33-t1/j-005ici3c33-t1_20221208T130227/logs/ Failure Details: ``` Message: subprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 seconds Type: None Text: self = <tests.manage.mcg.test_bucket_replication.TestReplication object at 0x7f91e590bb50> awscli_pod_session = <ocs_ci.ocs.resources.pod.Pod object at 0x7f91e61dca30> mcg_obj_session = <ocs_ci.ocs.resources.mcg.MCG object at 0x7f91e0e37910> bucket_factory = <function bucket_factory_fixture.<locals>._create_buckets at 0x7f91e42178b0> source_bucketclass = {'backingstore_dict': {'aws': [(1, 'eu-central-1')]}, 'interface': 'CLI'} target_bucketclass = {'backingstore_dict': {'azure': [(1, None)]}, 'interface': 'CLI'} @pytest.mark.parametrize( argnames=["source_bucketclass", "target_bucketclass"], argvalues=[ pytest.param( { "interface": "OC", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, {"interface": "OC", "backingstore_dict": {"azure": [(1, None)]}}, marks=[tier1, pytest.mark.polarion_id("OCS-2678")], ), pytest.param( { "interface": "OC", "backingstore_dict": {"gcp": [(1, None)]}, }, { "interface": "OC", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, marks=[tier2], ), pytest.param( { "interface": "CLI", "backingstore_dict": {"azure": [(1, None)]}, }, {"interface": "CLI", "backingstore_dict": {"gcp": [(1, None)]}}, marks=[tier2], ), pytest.param( { "interface": "CLI", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, {"interface": "CLI", "backingstore_dict": {"azure": [(1, None)]}}, marks=[tier1, pytest.mark.polarion_id("OCS-2679")], ), pytest.param( { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"aws": [(1, "eu-central-1")]}, }, }, { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"azure": [(1, None)]}, }, }, marks=[tier2], ), pytest.param( { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"azure": [(1, None)]}, }, }, { "interface": "CLI", "backingstore_dict": {"gcp": [(1, None)]}, }, marks=[tier1], ), ], ids=[ "AWStoAZURE-BS-OC", "GCPtoAWS-BS-OC", "AZUREtoCGP-BS-CLI", "AWStoAZURE-BS-CLI", "AWStoAZURE-NS-OC", "AZUREtoGCP-NS-Hybrid", ], ) def test_unidirectional_bucket_replication( self, awscli_pod_session, mcg_obj_session, bucket_factory, source_bucketclass, target_bucketclass, ): """ Test unidirectional bucket replication using CLI and YAML by adding objects to a backingstore- and namespacestore-backed buckets """ > target_bucket_name = bucket_factory(bucketclass=target_bucketclass)[0].name tests/manage/mcg/test_bucket_replication.py:121: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tests/conftest.py:2557: in _create_buckets bucketclass if bucketclass is None else bucket_class_factory(bucketclass) ocs_ci/ocs/resources/bucketclass.py:169: in _create_bucket_class for backingstore in backingstore_factory( ocs_ci/ocs/resources/backingstore.py:349: in _create_backingstore cmdMap[method.lower()][cloud.lower()]( ocs_ci/ocs/bucket_utils.py:602: in cli_create_azure_backingstore mcg_obj.exec_mcg_cmd( ocs_ci/ocs/resources/mcg.py:869: in exec_mcg_cmd result = exec_cmd( ocs_ci/utility/utils.py:607: in exec_cmd completed_process = subprocess.run( /usr/lib64/python3.8/subprocess.py:495: in run stdout, stderr = process.communicate(input, timeout=timeout) /usr/lib64/python3.8/subprocess.py:1028: in communicate stdout, stderr = self._communicate(input, endtime, timeout) /usr/lib64/python3.8/subprocess.py:1869: in _communicate self._check_timeout(endtime, orig_timeout, stdout, stderr) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <subprocess.Popen object at 0x7f920d985df0>, endtime = 10075.34677407 orig_timeout = 600, stdout_seq = [] stderr_seq = [b'time="2022-12-08T15:39:44Z" level=info msg="\xe2\x9c\x85 Exists: NooBaa \\"noobaa\\"\\n"\n', b'time="2022-12-08T15:...re-backingstore-8f648f21765242bd99c78\\" Phase is \\"Creating\\": TemporaryError Target bucket doesn\'t exist"\n', ...] skip_check_and_raise = False def _check_timeout(self, endtime, orig_timeout, stdout_seq, stderr_seq, skip_check_and_raise=False): """Convenience for checking if a timeout has expired.""" if endtime is None: return if skip_check_and_raise or _time() > endtime: > raise TimeoutExpired( self.args, orig_timeout, output=b''.join(stdout_seq) if stdout_seq else None, stderr=b''.join(stderr_seq) if stderr_seq else None) E subprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 seconds /usr/lib64/python3.8/subprocess.py:1072: TimeoutExpired ```
1.0
test_unidirectional_bucket_replication failed with ubprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 second - Run details: URL: https://reportportal-ocs4.apps.ocp-c1.prod.psi.redhat.com/ui/#OCS/launches/362/7364/320363/320416/320422/log Run ID: 1670509372 Test Case: test_unidirectional_bucket_replication ODF Build: 4.12.0-130 OCP Version: 4.12 Job name: IBM Cloud IPI 3AZ RHCOS 3M 3W tier1 Jenkins job: https://ocs4-jenkins-csb-odf-qe.apps.ocp-c1.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/6343/ Logs URL: http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j-005ici3c33-t1/j-005ici3c33-t1_20221208T130227/logs/ Failure Details: ``` Message: subprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 seconds Type: None Text: self = <tests.manage.mcg.test_bucket_replication.TestReplication object at 0x7f91e590bb50> awscli_pod_session = <ocs_ci.ocs.resources.pod.Pod object at 0x7f91e61dca30> mcg_obj_session = <ocs_ci.ocs.resources.mcg.MCG object at 0x7f91e0e37910> bucket_factory = <function bucket_factory_fixture.<locals>._create_buckets at 0x7f91e42178b0> source_bucketclass = {'backingstore_dict': {'aws': [(1, 'eu-central-1')]}, 'interface': 'CLI'} target_bucketclass = {'backingstore_dict': {'azure': [(1, None)]}, 'interface': 'CLI'} @pytest.mark.parametrize( argnames=["source_bucketclass", "target_bucketclass"], argvalues=[ pytest.param( { "interface": "OC", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, {"interface": "OC", "backingstore_dict": {"azure": [(1, None)]}}, marks=[tier1, pytest.mark.polarion_id("OCS-2678")], ), pytest.param( { "interface": "OC", "backingstore_dict": {"gcp": [(1, None)]}, }, { "interface": "OC", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, marks=[tier2], ), pytest.param( { "interface": "CLI", "backingstore_dict": {"azure": [(1, None)]}, }, {"interface": "CLI", "backingstore_dict": {"gcp": [(1, None)]}}, marks=[tier2], ), pytest.param( { "interface": "CLI", "backingstore_dict": {"aws": [(1, "eu-central-1")]}, }, {"interface": "CLI", "backingstore_dict": {"azure": [(1, None)]}}, marks=[tier1, pytest.mark.polarion_id("OCS-2679")], ), pytest.param( { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"aws": [(1, "eu-central-1")]}, }, }, { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"azure": [(1, None)]}, }, }, marks=[tier2], ), pytest.param( { "interface": "OC", "namespace_policy_dict": { "type": "Single", "namespacestore_dict": {"azure": [(1, None)]}, }, }, { "interface": "CLI", "backingstore_dict": {"gcp": [(1, None)]}, }, marks=[tier1], ), ], ids=[ "AWStoAZURE-BS-OC", "GCPtoAWS-BS-OC", "AZUREtoCGP-BS-CLI", "AWStoAZURE-BS-CLI", "AWStoAZURE-NS-OC", "AZUREtoGCP-NS-Hybrid", ], ) def test_unidirectional_bucket_replication( self, awscli_pod_session, mcg_obj_session, bucket_factory, source_bucketclass, target_bucketclass, ): """ Test unidirectional bucket replication using CLI and YAML by adding objects to a backingstore- and namespacestore-backed buckets """ > target_bucket_name = bucket_factory(bucketclass=target_bucketclass)[0].name tests/manage/mcg/test_bucket_replication.py:121: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tests/conftest.py:2557: in _create_buckets bucketclass if bucketclass is None else bucket_class_factory(bucketclass) ocs_ci/ocs/resources/bucketclass.py:169: in _create_bucket_class for backingstore in backingstore_factory( ocs_ci/ocs/resources/backingstore.py:349: in _create_backingstore cmdMap[method.lower()][cloud.lower()]( ocs_ci/ocs/bucket_utils.py:602: in cli_create_azure_backingstore mcg_obj.exec_mcg_cmd( ocs_ci/ocs/resources/mcg.py:869: in exec_mcg_cmd result = exec_cmd( ocs_ci/utility/utils.py:607: in exec_cmd completed_process = subprocess.run( /usr/lib64/python3.8/subprocess.py:495: in run stdout, stderr = process.communicate(input, timeout=timeout) /usr/lib64/python3.8/subprocess.py:1028: in communicate stdout, stderr = self._communicate(input, endtime, timeout) /usr/lib64/python3.8/subprocess.py:1869: in _communicate self._check_timeout(endtime, orig_timeout, stdout, stderr) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <subprocess.Popen object at 0x7f920d985df0>, endtime = 10075.34677407 orig_timeout = 600, stdout_seq = [] stderr_seq = [b'time="2022-12-08T15:39:44Z" level=info msg="\xe2\x9c\x85 Exists: NooBaa \\"noobaa\\"\\n"\n', b'time="2022-12-08T15:...re-backingstore-8f648f21765242bd99c78\\" Phase is \\"Creating\\": TemporaryError Target bucket doesn\'t exist"\n', ...] skip_check_and_raise = False def _check_timeout(self, endtime, orig_timeout, stdout_seq, stderr_seq, skip_check_and_raise=False): """Convenience for checking if a timeout has expired.""" if endtime is None: return if skip_check_and_raise or _time() > endtime: > raise TimeoutExpired( self.args, orig_timeout, output=b''.join(stdout_seq) if stdout_seq else None, stderr=b''.join(stderr_seq) if stderr_seq else None) E subprocess.TimeoutExpired: Command '['yes | /home/jenkins/workspace/qe-deploy-ocs-cluster-prod/ocs-ci/data/mcg-cli backingstore create azure-blob azure-backingstore-8f648f21765242bd99c78 --account-key ZoIEOXRWuTrGpRMHC0AOybbMEzycDjO35H8mWo0MmGJXcLZQOZmJWLSHvcbYs5PlnQb3iQO6BrebY/m7r1iYfQ== --account-name mcgtestingstorage --target-blob-container azure-uls-42da9b353ccb4fe6b87c27692a880e -n openshift-storage']' timed out after 600 seconds /usr/lib64/python3.8/subprocess.py:1072: TimeoutExpired ```
test
test unidirectional bucket replication failed with ubprocess timeoutexpired command timed out after second run details url run id test case test unidirectional bucket replication odf build ocp version job name ibm cloud ipi rhcos jenkins job logs url failure details message subprocess timeoutexpired command timed out after seconds type none text self awscli pod session mcg obj session bucket factory create buckets at source bucketclass backingstore dict aws interface cli target bucketclass backingstore dict azure interface cli pytest mark parametrize argnames argvalues pytest param interface oc backingstore dict aws interface oc backingstore dict azure marks pytest param interface oc backingstore dict gcp interface oc backingstore dict aws marks pytest param interface cli backingstore dict azure interface cli backingstore dict gcp marks pytest param interface cli backingstore dict aws interface cli backingstore dict azure marks pytest param interface oc namespace policy dict type single namespacestore dict aws interface oc namespace policy dict type single namespacestore dict azure marks pytest param interface oc namespace policy dict type single namespacestore dict azure interface cli backingstore dict gcp marks ids awstoazure bs oc gcptoaws bs oc azuretocgp bs cli awstoazure bs cli awstoazure ns oc azuretogcp ns hybrid def test unidirectional bucket replication self awscli pod session mcg obj session bucket factory source bucketclass target bucketclass test unidirectional bucket replication using cli and yaml by adding objects to a backingstore and namespacestore backed buckets target bucket name bucket factory bucketclass target bucketclass name tests manage mcg test bucket replication py tests conftest py in create buckets bucketclass if bucketclass is none else bucket class factory bucketclass ocs ci ocs resources bucketclass py in create bucket class for backingstore in backingstore factory ocs ci ocs resources backingstore py in create backingstore cmdmap ocs ci ocs bucket utils py in cli create azure backingstore mcg obj exec mcg cmd ocs ci ocs resources mcg py in exec mcg cmd result exec cmd ocs ci utility utils py in exec cmd completed process subprocess run usr subprocess py in run stdout stderr process communicate input timeout timeout usr subprocess py in communicate stdout stderr self communicate input endtime timeout usr subprocess py in communicate self check timeout endtime orig timeout stdout stderr self endtime orig timeout stdout seq stderr seq skip check and raise false def check timeout self endtime orig timeout stdout seq stderr seq skip check and raise false convenience for checking if a timeout has expired if endtime is none return if skip check and raise or time endtime raise timeoutexpired self args orig timeout output b join stdout seq if stdout seq else none stderr b join stderr seq if stderr seq else none e subprocess timeoutexpired command timed out after seconds usr subprocess py timeoutexpired
1
130,820
10,660,942,945
IssuesEvent
2019-10-18 11:06:53
elastic/logstash
https://api.github.com/repos/elastic/logstash
opened
Test failure: "Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents"
test failure
``` 11:51:00 Failures: 11:51:00 11:51:00 1) Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents 11:51:00 Failure/Error: Unable to find matching line from backtrace 11:51:00 11:51:00 expected: 1000 11:51:00 got: 962 11:51:00 11:51:00 (compared using ==) 11:51:00 Shared Example Group: "it can send 1000 documents to and index from the dlq" called from ./specs/dlq_spec.rb:93 11:51:00 11:51:00 Finished in 20 minutes 33 seconds (files took 7.85 seconds to load) 11:51:00 25 examples, 1 failure, 1 pending 11:51:00 11:51:00 Failed examples: 11:51:00 11:51:00 rspec './specs/dlq_spec.rb[1:1:1:1:1]' # Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents 11:51:00 11:51:00 Randomized with seed 46841 ```
1.0
Test failure: "Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents" - ``` 11:51:00 Failures: 11:51:00 11:51:00 1) Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents 11:51:00 Failure/Error: Unable to find matching line from backtrace 11:51:00 11:51:00 expected: 1000 11:51:00 got: 962 11:51:00 11:51:00 (compared using ==) 11:51:00 Shared Example Group: "it can send 1000 documents to and index from the dlq" called from ./specs/dlq_spec.rb:93 11:51:00 11:51:00 Finished in 20 minutes 33 seconds (files took 7.85 seconds to load) 11:51:00 25 examples, 1 failure, 1 pending 11:51:00 11:51:00 Failed examples: 11:51:00 11:51:00 rspec './specs/dlq_spec.rb[1:1:1:1:1]' # Test Dead Letter Queue using pipelines.yml with multiple pipelines behaves like it can send 1000 documents to and index from the dlq should index all documents 11:51:00 11:51:00 Randomized with seed 46841 ```
test
test failure test dead letter queue using pipelines yml with multiple pipelines behaves like it can send documents to and index from the dlq should index all documents failures test dead letter queue using pipelines yml with multiple pipelines behaves like it can send documents to and index from the dlq should index all documents failure error unable to find matching line from backtrace expected got compared using shared example group it can send documents to and index from the dlq called from specs dlq spec rb finished in minutes seconds files took seconds to load examples failure pending failed examples rspec specs dlq spec rb test dead letter queue using pipelines yml with multiple pipelines behaves like it can send documents to and index from the dlq should index all documents randomized with seed
1
511,215
14,856,338,847
IssuesEvent
2021-01-18 14:01:41
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Gradle dependency Version using variable and set()
manager:gradle manager:gradle-lite priority-3-normal type:feature
**What Renovate type, platform and version are you using?** we are using https://gitlab.com/renovate-bot/renovate-runner/-/blob/master/templates/renovate.gitlab-ci.yml as an include in gitlab-ci **Describe the bug** When generating a sample project at start.pring.io for gradle im getting this (I already changed BOM as found in #4614 to platform()): ``` plugins { id 'org.springframework.boot' version '2.4.1' id 'io.spring.dependency-management' version '1.0.10.RELEASE' id 'java' } group = 'com.example' version = '0.0.1-SNAPSHOT' sourceCompatibility = '11' repositories { mavenCentral() } ext { set('springCloudVersion', "Hoxton.SR8") } dependencies { implementation platform("org.springframework.cloud:spring-cloud-dependencies:${springCloudVersion}") implementation 'org.springframework.cloud:spring-cloud-starter-sleuth' testImplementation 'org.springframework.boot:spring-boot-starter-test' } test { useJUnitPlatform() } ``` Renovate fails to upgrade ` set('springCloudVersion', "Hoxton.SR8")`, using `springCloudVersion = "Hoxton.SR8"` it does work. regex to match this : `set\(\s*['"](.+)['"]\s*,\s*['"](.+)['"]\s*\)` and replacement: `$1 = "$2"` Is there any chance to get this gradle syntax supported? **Relevant debug logs** ``` DEBUG: gradle.updateDependency(): packageFile:settings.gradle depName:org.springframework.cloud:spring-cloud-dependencies, version:Hoxton.SR8 ==> 2020.0.0 (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: gradle.updateDependency(): packageFile:build.gradle depName:org.springframework.cloud:spring-cloud-dependencies, version:Hoxton.SR8 ==> 2020.0.0 (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No package files need updating (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No updated lock files in branch (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No files to commit (repository=23671976, branch=renovate/major-spring-cloud) ``` **Have you created a minimal reproduction repository?** https://gitlab.com/vgropp/renovate-spring-gradle-sample
1.0
Gradle dependency Version using variable and set() - **What Renovate type, platform and version are you using?** we are using https://gitlab.com/renovate-bot/renovate-runner/-/blob/master/templates/renovate.gitlab-ci.yml as an include in gitlab-ci **Describe the bug** When generating a sample project at start.pring.io for gradle im getting this (I already changed BOM as found in #4614 to platform()): ``` plugins { id 'org.springframework.boot' version '2.4.1' id 'io.spring.dependency-management' version '1.0.10.RELEASE' id 'java' } group = 'com.example' version = '0.0.1-SNAPSHOT' sourceCompatibility = '11' repositories { mavenCentral() } ext { set('springCloudVersion', "Hoxton.SR8") } dependencies { implementation platform("org.springframework.cloud:spring-cloud-dependencies:${springCloudVersion}") implementation 'org.springframework.cloud:spring-cloud-starter-sleuth' testImplementation 'org.springframework.boot:spring-boot-starter-test' } test { useJUnitPlatform() } ``` Renovate fails to upgrade ` set('springCloudVersion', "Hoxton.SR8")`, using `springCloudVersion = "Hoxton.SR8"` it does work. regex to match this : `set\(\s*['"](.+)['"]\s*,\s*['"](.+)['"]\s*\)` and replacement: `$1 = "$2"` Is there any chance to get this gradle syntax supported? **Relevant debug logs** ``` DEBUG: gradle.updateDependency(): packageFile:settings.gradle depName:org.springframework.cloud:spring-cloud-dependencies, version:Hoxton.SR8 ==> 2020.0.0 (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: gradle.updateDependency(): packageFile:build.gradle depName:org.springframework.cloud:spring-cloud-dependencies, version:Hoxton.SR8 ==> 2020.0.0 (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No package files need updating (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No updated lock files in branch (repository=23671976, branch=renovate/major-spring-cloud) DEBUG: No files to commit (repository=23671976, branch=renovate/major-spring-cloud) ``` **Have you created a minimal reproduction repository?** https://gitlab.com/vgropp/renovate-spring-gradle-sample
non_test
gradle dependency version using variable and set what renovate type platform and version are you using we are using as an include in gitlab ci describe the bug when generating a sample project at start pring io for gradle im getting this i already changed bom as found in to platform plugins id org springframework boot version id io spring dependency management version release id java group com example version snapshot sourcecompatibility repositories mavencentral ext set springcloudversion hoxton dependencies implementation platform org springframework cloud spring cloud dependencies springcloudversion implementation org springframework cloud spring cloud starter sleuth testimplementation org springframework boot spring boot starter test test usejunitplatform renovate fails to upgrade set springcloudversion hoxton using springcloudversion hoxton it does work regex to match this set s s s s and replacement is there any chance to get this gradle syntax supported relevant debug logs debug gradle updatedependency packagefile settings gradle depname org springframework cloud spring cloud dependencies version hoxton repository branch renovate major spring cloud debug gradle updatedependency packagefile build gradle depname org springframework cloud spring cloud dependencies version hoxton repository branch renovate major spring cloud debug no package files need updating repository branch renovate major spring cloud debug no updated lock files in branch repository branch renovate major spring cloud debug no files to commit repository branch renovate major spring cloud have you created a minimal reproduction repository
0
230,374
18,546,606,571
IssuesEvent
2021-10-21 23:28:10
dotnet/machinelearning-modelbuilder
https://api.github.com/repos/dotnet/machinelearning-modelbuilder
closed
Squiggles are not displayed for the code in errors in Notebook file.
Priority:1 Test Team Notebook
**System Information (please complete the following information):** - Microsoft Visual Studio Enterprise 2022 Preview: 17.0.0 Preview 5.0 - ML.Net Model Builder: 16.8.3.2151403 (Main Branch) - Notebook Editor [Preview]: 0.2.1.2151402 **Describe the bug** - On which step of the process did you run into an issue: The problematic code in the cell. **To Reproduce** Steps to reproduce the behavior: 1. Select Create a new project from the Visual Studio 2022 start window; 2. Choose the C# Console App (.NET Core) project template with .Net 5.0; 3. Right click the ConsoleApp project>Add>New Item...>Notebook; 4. Write invalid code in the code cell, see that Squiggles are not displayed under the code. **Expected behavior** Display squiggles for the code in errors in Notebook file. **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/81727020/137457928-1e8a9b70-c82b-4d53-b852-32f563eee7f7.png) **Additional context** Work in VS Code. ![image](https://user-images.githubusercontent.com/81727020/137457983-45135b04-a519-479c-969b-0db986504973.png)
1.0
Squiggles are not displayed for the code in errors in Notebook file. - **System Information (please complete the following information):** - Microsoft Visual Studio Enterprise 2022 Preview: 17.0.0 Preview 5.0 - ML.Net Model Builder: 16.8.3.2151403 (Main Branch) - Notebook Editor [Preview]: 0.2.1.2151402 **Describe the bug** - On which step of the process did you run into an issue: The problematic code in the cell. **To Reproduce** Steps to reproduce the behavior: 1. Select Create a new project from the Visual Studio 2022 start window; 2. Choose the C# Console App (.NET Core) project template with .Net 5.0; 3. Right click the ConsoleApp project>Add>New Item...>Notebook; 4. Write invalid code in the code cell, see that Squiggles are not displayed under the code. **Expected behavior** Display squiggles for the code in errors in Notebook file. **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/81727020/137457928-1e8a9b70-c82b-4d53-b852-32f563eee7f7.png) **Additional context** Work in VS Code. ![image](https://user-images.githubusercontent.com/81727020/137457983-45135b04-a519-479c-969b-0db986504973.png)
test
squiggles are not displayed for the code in errors in notebook file system information please complete the following information microsoft visual studio enterprise preview preview ml net model builder main branch notebook editor describe the bug on which step of the process did you run into an issue the problematic code in the cell to reproduce steps to reproduce the behavior select create a new project from the visual studio start window choose the c console app net core project template with net right click the consoleapp project add new item notebook write invalid code in the code cell see that squiggles are not displayed under the code expected behavior display squiggles for the code in errors in notebook file screenshots if applicable add screenshots to help explain your problem additional context work in vs code
1
49,974
6,288,950,632
IssuesEvent
2017-07-19 18:07:28
roschaefer/story.board
https://api.github.com/repos/roschaefer/story.board
closed
tc Channel select: Sensorstory as a default
design Priority: medium User Story
As a reporter I want sensorstory as the default channel in order to save one click.
1.0
tc Channel select: Sensorstory as a default - As a reporter I want sensorstory as the default channel in order to save one click.
non_test
tc channel select sensorstory as a default as a reporter i want sensorstory as the default channel in order to save one click
0
135,114
10,962,867,554
IssuesEvent
2019-11-27 18:15:26
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
Chromium links are shown on toptiles on clean launch
QA/Test-Plan-Specified QA/Yes android-core-release
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Chromium links are shown on top tiles on clean launch ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 2.0.0 2. Skip onboarding flow 3. Chromium links are shown on top tiles ## Actual result: <!--Please add screenshots if needed--> Chromium links are shown on top tiles on clean launch ![image](https://user-images.githubusercontent.com/38657976/69749077-b5f51000-116f-11ea-881b-4aefef196497.png) ## Expected result: Chromium links shouldn't be shown on top tiles on clean launch ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave 2.0.0 Chromium 78.1.3.15 ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @emerick @yrliou @SergeyZhukovsky
1.0
Chromium links are shown on toptiles on clean launch - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Chromium links are shown on top tiles on clean launch ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 2.0.0 2. Skip onboarding flow 3. Chromium links are shown on top tiles ## Actual result: <!--Please add screenshots if needed--> Chromium links are shown on top tiles on clean launch ![image](https://user-images.githubusercontent.com/38657976/69749077-b5f51000-116f-11ea-881b-4aefef196497.png) ## Expected result: Chromium links shouldn't be shown on top tiles on clean launch ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave 2.0.0 Chromium 78.1.3.15 ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @emerick @yrliou @SergeyZhukovsky
test
chromium links are shown on toptiles on clean launch have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description chromium links are shown on top tiles on clean launch steps to reproduce clean profile skip onboarding flow chromium links are shown on top tiles actual result chromium links are shown on top tiles on clean launch expected result chromium links shouldn t be shown on top tiles on clean launch reproduces how often always brave version brave version info brave chromium version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel no can you reproduce this issue with the dev channel no can you reproduce this issue with the nightly channel no other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa emerick yrliou sergeyzhukovsky
1
48,328
5,954,082,896
IssuesEvent
2017-05-27 14:30:11
lolmos/statictestting
https://api.github.com/repos/lolmos/statictestting
closed
Make all activities content and nav live
Activities section (1) Test
- [ ] Global nav changes (both languages) - [ ] Ensure tabs are live (both languages) - [ ] Ensure three sections and content pages are live (both languages) - [ ] Test PDF links - [ ] Check nothing else has broken in the process
1.0
Make all activities content and nav live - - [ ] Global nav changes (both languages) - [ ] Ensure tabs are live (both languages) - [ ] Ensure three sections and content pages are live (both languages) - [ ] Test PDF links - [ ] Check nothing else has broken in the process
test
make all activities content and nav live global nav changes both languages ensure tabs are live both languages ensure three sections and content pages are live both languages test pdf links check nothing else has broken in the process
1
771,649
27,088,150,364
IssuesEvent
2023-02-14 18:37:41
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
[Nightly CI Failures] Failures detected for google-cloud-spanner-v1
type: bug priority: p1 nightly failure
At 2023-02-09 08:59:47 UTC, detected failures in google-cloud-spanner-v1 for: rubocop report_key_ffba2d5871481bc08dc1434eb0d067c2
1.0
[Nightly CI Failures] Failures detected for google-cloud-spanner-v1 - At 2023-02-09 08:59:47 UTC, detected failures in google-cloud-spanner-v1 for: rubocop report_key_ffba2d5871481bc08dc1434eb0d067c2
non_test
failures detected for google cloud spanner at utc detected failures in google cloud spanner for rubocop report key
0
2,134
3,520,718,463
IssuesEvent
2016-01-12 22:02:10
twosigma/beaker-notebook
https://api.github.com/repos/twosigma/beaker-notebook
closed
add option for location of prefs and temp files
Enhancement Infrastructure Priority High
this is especially important on windows in order to make a portable version of beaker that you can run from a thumb drive and does not modify any other volume. this would be a cmd line, like --plugin-option and --public-server
1.0
add option for location of prefs and temp files - this is especially important on windows in order to make a portable version of beaker that you can run from a thumb drive and does not modify any other volume. this would be a cmd line, like --plugin-option and --public-server
non_test
add option for location of prefs and temp files this is especially important on windows in order to make a portable version of beaker that you can run from a thumb drive and does not modify any other volume this would be a cmd line like plugin option and public server
0
135,325
10,974,771,631
IssuesEvent
2019-11-29 09:54:16
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Smoke test failure: Debug stepOver, stepIn, stepOut
bug debug smoke-test
Insider version: Version: 1.40.0-insider Commit: c8846de84ab2f0a9279a41d80414465124c52a7a Date: 2019-11-01T05:35:06.002Z Electron: 6.1.2 Chrome: 76.0.3809.146 Node.js: 12.4.0 V8: 7.6.303.31-electron.0 OS: Linux x64 5.3.0-19-generic Running Code Debug stepOver, stepIn, stepOut: Error: Timeout: get elements 'div[id="workbench.view.debug"] .monaco-list-row .stack-frame' after 20 seconds. at /home/dirkb/Projects/mseng/VSCode/vscode/test/automation/out/code.js:175:23 at Generator.next (<anonymous>) at fulfilled (/home/dirkb/Projects/mseng/VSCode/vscode/test/automation/out/code.js:8:58)
1.0
Smoke test failure: Debug stepOver, stepIn, stepOut - Insider version: Version: 1.40.0-insider Commit: c8846de84ab2f0a9279a41d80414465124c52a7a Date: 2019-11-01T05:35:06.002Z Electron: 6.1.2 Chrome: 76.0.3809.146 Node.js: 12.4.0 V8: 7.6.303.31-electron.0 OS: Linux x64 5.3.0-19-generic Running Code Debug stepOver, stepIn, stepOut: Error: Timeout: get elements 'div[id="workbench.view.debug"] .monaco-list-row .stack-frame' after 20 seconds. at /home/dirkb/Projects/mseng/VSCode/vscode/test/automation/out/code.js:175:23 at Generator.next (<anonymous>) at fulfilled (/home/dirkb/Projects/mseng/VSCode/vscode/test/automation/out/code.js:8:58)
test
smoke test failure debug stepover stepin stepout insider version version insider commit date electron chrome node js electron os linux generic running code debug stepover stepin stepout error timeout get elements div monaco list row stack frame after seconds at home dirkb projects mseng vscode vscode test automation out code js at generator next at fulfilled home dirkb projects mseng vscode vscode test automation out code js
1
506,399
14,664,253,821
IssuesEvent
2020-12-29 11:33:33
bryntum/support
https://api.github.com/repos/bryntum/support
closed
Zooming in configuration demo breaks time axis
bug forum high-priority resolved
[Forum post](https://www.bryntum.com/forum/viewtopic.php?f=44&t=16217&p=80755) https://www.bryntum.com/examples/scheduler/configuration/ If we run the above example(configuration), i found the time axis has left-over values Steps Click Zoom In Check the time axis ![Zoom-in Defect](https://www.bryntum.com/forum/download/file.php?id=12044) If we scroll ahead it's seen that, it continues elsewhere too ![Zoom-in Defect](https://www.bryntum.com/forum/download/file.php?id=12044)
1.0
Zooming in configuration demo breaks time axis - [Forum post](https://www.bryntum.com/forum/viewtopic.php?f=44&t=16217&p=80755) https://www.bryntum.com/examples/scheduler/configuration/ If we run the above example(configuration), i found the time axis has left-over values Steps Click Zoom In Check the time axis ![Zoom-in Defect](https://www.bryntum.com/forum/download/file.php?id=12044) If we scroll ahead it's seen that, it continues elsewhere too ![Zoom-in Defect](https://www.bryntum.com/forum/download/file.php?id=12044)
non_test
zooming in configuration demo breaks time axis if we run the above example configuration i found the time axis has left over values steps click zoom in check the time axis if we scroll ahead it s seen that it continues elsewhere too
0
133,190
10,799,399,573
IssuesEvent
2019-11-06 12:06:24
bbc/simorgh
https://api.github.com/repos/bbc/simorgh
closed
Use more reliable selectors in AV e2e tests
Refinement Needed articles-av-epic articles-features-stream integration-testing ws-media-stream
**Is your feature request related to a problem? Please describe.** We need a more robust way of selecting DOM nodes in our AV e2e tests. Using class selectors, when we don't control the the classes (Styled Components does), is fragile and sets us up for breakage. **Describe the solution you'd like** Explore the use of data attributes - there is already a precedence for `data-e2e=""` - and apply these to the relevant AV components. This will need a child Psammead issue. **Describe alternatives you've considered** N/A **Testing notes** [Tester to complete] Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc - [ ] This feature is expected to need manual testing. **Additional context** Add any other context or screenshots about the feature request here.
1.0
Use more reliable selectors in AV e2e tests - **Is your feature request related to a problem? Please describe.** We need a more robust way of selecting DOM nodes in our AV e2e tests. Using class selectors, when we don't control the the classes (Styled Components does), is fragile and sets us up for breakage. **Describe the solution you'd like** Explore the use of data attributes - there is already a precedence for `data-e2e=""` - and apply these to the relevant AV components. This will need a child Psammead issue. **Describe alternatives you've considered** N/A **Testing notes** [Tester to complete] Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc - [ ] This feature is expected to need manual testing. **Additional context** Add any other context or screenshots about the feature request here.
test
use more reliable selectors in av tests is your feature request related to a problem please describe we need a more robust way of selecting dom nodes in our av tests using class selectors when we don t control the the classes styled components does is fragile and sets us up for breakage describe the solution you d like explore the use of data attributes there is already a precedence for data and apply these to the relevant av components this will need a child psammead issue describe alternatives you ve considered n a testing notes dev insight will cypress tests be required or are unit tests sufficient will there be any potential regression etc this feature is expected to need manual testing additional context add any other context or screenshots about the feature request here
1
222,293
24,697,508,856
IssuesEvent
2022-10-19 13:11:10
TreyM-WSS/ol-kit
https://api.github.com/repos/TreyM-WSS/ol-kit
opened
CVE-2022-37601 (High) detected in loader-utils-1.4.0.tgz
security vulnerability
## CVE-2022-37601 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-8.0.6.tgz (Root Library) - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function parseQuery in parseQuery.js in webpack loader-utils 2.0.0 via the name variable in parseQuery.js. <p>Publish Date: 2022-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37601>CVE-2022-37601</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-12</p> <p>Fix Resolution (loader-utils): 2.0.0</p> <p>Direct dependency fix Resolution (babel-loader): 8.2.4</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2022-37601 (High) detected in loader-utils-1.4.0.tgz - ## CVE-2022-37601 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-8.0.6.tgz (Root Library) - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function parseQuery in parseQuery.js in webpack loader-utils 2.0.0 via the name variable in parseQuery.js. <p>Publish Date: 2022-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37601>CVE-2022-37601</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-12</p> <p>Fix Resolution (loader-utils): 2.0.0</p> <p>Direct dependency fix Resolution (babel-loader): 8.2.4</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_test
cve high detected in loader utils tgz cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy babel loader tgz root library x loader utils tgz vulnerable library found in base branch master vulnerability details prototype pollution vulnerability in function parsequery in parsequery js in webpack loader utils via the name variable in parsequery js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution loader utils direct dependency fix resolution babel loader rescue worker helmet automatic remediation is available for this issue
0
307,770
26,561,698,992
IssuesEvent
2023-01-20 16:22:51
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
closed
Project items window
type - enhancement f - computations f - API f - inspector feature f - testing
This is a debugging/information tool for listing all internal objects in a minimal UI fashion. This is almost required to move forward on more complex computations and data structures used in upcoming applications (such as elemental analysis). - [x] nion-software/nionui#66 - [x] Implement default component handler registration in base Declarative.Handler - [ ] List all referrers/containers of each component type - [x] Register all component types and allow browsing - [x] Allow component types to register a "best representation" handler - [ ] Implement list/map/set property handlers - [ ] Implement back/forward buttons in browser - [ ] Implement tabs in project inspector for multiple paths in browser
1.0
Project items window - This is a debugging/information tool for listing all internal objects in a minimal UI fashion. This is almost required to move forward on more complex computations and data structures used in upcoming applications (such as elemental analysis). - [x] nion-software/nionui#66 - [x] Implement default component handler registration in base Declarative.Handler - [ ] List all referrers/containers of each component type - [x] Register all component types and allow browsing - [x] Allow component types to register a "best representation" handler - [ ] Implement list/map/set property handlers - [ ] Implement back/forward buttons in browser - [ ] Implement tabs in project inspector for multiple paths in browser
test
project items window this is a debugging information tool for listing all internal objects in a minimal ui fashion this is almost required to move forward on more complex computations and data structures used in upcoming applications such as elemental analysis nion software nionui implement default component handler registration in base declarative handler list all referrers containers of each component type register all component types and allow browsing allow component types to register a best representation handler implement list map set property handlers implement back forward buttons in browser implement tabs in project inspector for multiple paths in browser
1