Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
23,388
16,106,047,490
IssuesEvent
2021-04-27 15:02:41
emory-libraries/blacklight-catalog
https://api.github.com/repos/emory-libraries/blacklight-catalog
closed
Make solr node cpu sizes consistent
Infrastructure Story
**ISSUE** We currently have different EC2 instances configured for the Solr Cloud nodes in the Test environment compared to Arch and Prod environments. We've decided it makes more sense for now to have all environments identically sized. **ACCEPTANCE** Each environment's Solr Cloud is configured with 3 nodes with each node running on a T3.Large instance - [x] ARCH (probably done already but triple check) - [x] TEST - [x] PROD (probably done already but triple check)
1.0
Make solr node cpu sizes consistent - **ISSUE** We currently have different EC2 instances configured for the Solr Cloud nodes in the Test environment compared to Arch and Prod environments. We've decided it makes more sense for now to have all environments identically sized. **ACCEPTANCE** Each environment's Solr Cloud is configured with 3 nodes with each node running on a T3.Large instance - [x] ARCH (probably done already but triple check) - [x] TEST - [x] PROD (probably done already but triple check)
infrastructure
make solr node cpu sizes consistent issue we currently have different instances configured for the solr cloud nodes in the test environment compared to arch and prod environments we ve decided it makes more sense for now to have all environments identically sized acceptance each environment s solr cloud is configured with nodes with each node running on a large instance arch probably done already but triple check test prod probably done already but triple check
1
12,925
10,022,082,493
IssuesEvent
2019-07-16 15:53:00
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
Generate .nupkg files that contain aspnetcore-runtime MSI for insertion into Visual Studio
area-infrastructure
We are working with the SDK team to simplifying the insertion of the .NET Core SDK into Visual Studio. https://github.com/dotnet/core-sdk/issues/2166 Today, the ASP.NET Core runtime is inserted into VS via the .NET Core SDK. As a part of breaking the SDK into smaller pieces, we need to add a new build step to 2.1/2.2 and 3.0 builds which produce the following NuGet packages. * AspNetCore.SharedFramework.$(Arch).$(MajorVersion).$(MinorVersion) * ID = VS.Redist.Common.AspNetCore.SharedFramework.$(Arch).$(MajorVersion).$(MinorVersion) * Contents * $pkgRoot/aspnetcore-runtime-$(version)-win-$(Arch).msi * $pkgRoot/sfx_$(Arch).cab * Versions: * 2.1 * 2.2 * 3.0 * Arch * x86 * x64 * AspNetCore.TargetingPack.$(Arch).$(MajorVersion).$(MinorVersion) * ID = VS.Redist.Common.AspNetCore.TargetingPack.$(Arch).$(MajorVersion).$(MinorVersion) * Contents * $pkgRoot/aspnetcore-targetingpack-$(version)-win-$(Arch).msi * Versions: * 3.0 * Arch * x86 * x64 ### Implementation suggestions The simplest implementation can use `nuget.exe` and a .nuspec file to wrap the files in a nupkg. We don't need to code-sign the nupkg or ship it to nuget.org Example: https://github.com/dotnet/core-sdk/blob/844978f4c2a042abf5fc924729466a4dce3c047d/src/redist/targets/packaging/windows/clisdk/VS.Redist.Common.NetCore.Toolset.nuspec /cc @mmitche @johnbeisner
1.0
Generate .nupkg files that contain aspnetcore-runtime MSI for insertion into Visual Studio - We are working with the SDK team to simplifying the insertion of the .NET Core SDK into Visual Studio. https://github.com/dotnet/core-sdk/issues/2166 Today, the ASP.NET Core runtime is inserted into VS via the .NET Core SDK. As a part of breaking the SDK into smaller pieces, we need to add a new build step to 2.1/2.2 and 3.0 builds which produce the following NuGet packages. * AspNetCore.SharedFramework.$(Arch).$(MajorVersion).$(MinorVersion) * ID = VS.Redist.Common.AspNetCore.SharedFramework.$(Arch).$(MajorVersion).$(MinorVersion) * Contents * $pkgRoot/aspnetcore-runtime-$(version)-win-$(Arch).msi * $pkgRoot/sfx_$(Arch).cab * Versions: * 2.1 * 2.2 * 3.0 * Arch * x86 * x64 * AspNetCore.TargetingPack.$(Arch).$(MajorVersion).$(MinorVersion) * ID = VS.Redist.Common.AspNetCore.TargetingPack.$(Arch).$(MajorVersion).$(MinorVersion) * Contents * $pkgRoot/aspnetcore-targetingpack-$(version)-win-$(Arch).msi * Versions: * 3.0 * Arch * x86 * x64 ### Implementation suggestions The simplest implementation can use `nuget.exe` and a .nuspec file to wrap the files in a nupkg. We don't need to code-sign the nupkg or ship it to nuget.org Example: https://github.com/dotnet/core-sdk/blob/844978f4c2a042abf5fc924729466a4dce3c047d/src/redist/targets/packaging/windows/clisdk/VS.Redist.Common.NetCore.Toolset.nuspec /cc @mmitche @johnbeisner
infrastructure
generate nupkg files that contain aspnetcore runtime msi for insertion into visual studio we are working with the sdk team to simplifying the insertion of the net core sdk into visual studio today the asp net core runtime is inserted into vs via the net core sdk as a part of breaking the sdk into smaller pieces we need to add a new build step to and builds which produce the following nuget packages aspnetcore sharedframework arch majorversion minorversion id vs redist common aspnetcore sharedframework arch majorversion minorversion contents pkgroot aspnetcore runtime version win arch msi pkgroot sfx arch cab versions arch aspnetcore targetingpack arch majorversion minorversion id vs redist common aspnetcore targetingpack arch majorversion minorversion contents pkgroot aspnetcore targetingpack version win arch msi versions arch implementation suggestions the simplest implementation can use nuget exe and a nuspec file to wrap the files in a nupkg we don t need to code sign the nupkg or ship it to nuget org example cc mmitche johnbeisner
1
2,158
3,535,404,910
IssuesEvent
2016-01-16 13:47:20
asciidoctor/asciidoctor.js
https://api.github.com/repos/asciidoctor/asciidoctor.js
closed
Travis CI - Test should fail
infrastructure
Test should fail because Opal is undefined but the job pass ![image](https://cloud.githubusercontent.com/assets/2006548/10366016/d4ccfbbe-6dc7-11e5-9904-d5e61578e07c.png) https://travis-ci.org/asciidoctor/asciidoctor.js/builds/83552848
1.0
Travis CI - Test should fail - Test should fail because Opal is undefined but the job pass ![image](https://cloud.githubusercontent.com/assets/2006548/10366016/d4ccfbbe-6dc7-11e5-9904-d5e61578e07c.png) https://travis-ci.org/asciidoctor/asciidoctor.js/builds/83552848
infrastructure
travis ci test should fail test should fail because opal is undefined but the job pass
1
16,741
12,129,547,898
IssuesEvent
2020-04-22 22:53:16
Xilinx/finn
https://api.github.com/repos/Xilinx/finn
closed
Support parallelizing node-local transformations
enhancement infrastructure
Several FINN transformations like Vivado HLS IPGen and npysim compilation take a long time to run, which slows down development cycles. These transformations could be sped up considerably by running several nodes in parallel. Although parallelizing general-purpose FINN graph transformations is a bit tricky due to potential sync issues, Vivado HLS IPGen etc. are all node-local transformations, i.e. the transform only accesses attributes from a single node, and modifies attributes on the same node as a result. These transformations are a lot easier to parallelize. Here's a small proposal for getting there: - Introduce a new kind of `Transformation` subclass called `NodeLocalTransformation`, with a new abstract class method called ´applyNodeLocal`. - This `applyNodeLocal` should take in a single ONNX node, and return a (possibly modified) single ONNX node. It should not have access to the entire `model` like the standard `Transformation`s do. - The `apply` member of `NodeLocalTransformation` is responsible for parallel launching of tasks, something like: ``` pool = multiprocessing.Pool(NUM_DEFAULT_WORKERS) old_nodes = model.graph.node new_nodes = list(*pool.map(lambda x: self.applyNodeLocal(x), old_nodes)) model.graph.node = new_nodes ``` Once the basic infrastructure in place, we can start changing existing `Transformation` derivatives to be `NodeLocalTransformation` derivatives instead. Here is a (probaby incomplete) list of node-local transformations, most compute-heavy ones are marked with +: ``` CleanUp Compile ++ HLSSynth_IPGen +++ ReplaceVerilogRelPaths SetExecMode ```
1.0
Support parallelizing node-local transformations - Several FINN transformations like Vivado HLS IPGen and npysim compilation take a long time to run, which slows down development cycles. These transformations could be sped up considerably by running several nodes in parallel. Although parallelizing general-purpose FINN graph transformations is a bit tricky due to potential sync issues, Vivado HLS IPGen etc. are all node-local transformations, i.e. the transform only accesses attributes from a single node, and modifies attributes on the same node as a result. These transformations are a lot easier to parallelize. Here's a small proposal for getting there: - Introduce a new kind of `Transformation` subclass called `NodeLocalTransformation`, with a new abstract class method called ´applyNodeLocal`. - This `applyNodeLocal` should take in a single ONNX node, and return a (possibly modified) single ONNX node. It should not have access to the entire `model` like the standard `Transformation`s do. - The `apply` member of `NodeLocalTransformation` is responsible for parallel launching of tasks, something like: ``` pool = multiprocessing.Pool(NUM_DEFAULT_WORKERS) old_nodes = model.graph.node new_nodes = list(*pool.map(lambda x: self.applyNodeLocal(x), old_nodes)) model.graph.node = new_nodes ``` Once the basic infrastructure in place, we can start changing existing `Transformation` derivatives to be `NodeLocalTransformation` derivatives instead. Here is a (probaby incomplete) list of node-local transformations, most compute-heavy ones are marked with +: ``` CleanUp Compile ++ HLSSynth_IPGen +++ ReplaceVerilogRelPaths SetExecMode ```
infrastructure
support parallelizing node local transformations several finn transformations like vivado hls ipgen and npysim compilation take a long time to run which slows down development cycles these transformations could be sped up considerably by running several nodes in parallel although parallelizing general purpose finn graph transformations is a bit tricky due to potential sync issues vivado hls ipgen etc are all node local transformations i e the transform only accesses attributes from a single node and modifies attributes on the same node as a result these transformations are a lot easier to parallelize here s a small proposal for getting there introduce a new kind of transformation subclass called nodelocaltransformation with a new abstract class method called ´applynodelocal this applynodelocal should take in a single onnx node and return a possibly modified single onnx node it should not have access to the entire model like the standard transformation s do the apply member of nodelocaltransformation is responsible for parallel launching of tasks something like pool multiprocessing pool num default workers old nodes model graph node new nodes list pool map lambda x self applynodelocal x old nodes model graph node new nodes once the basic infrastructure in place we can start changing existing transformation derivatives to be nodelocaltransformation derivatives instead here is a probaby incomplete list of node local transformations most compute heavy ones are marked with cleanup compile hlssynth ipgen replaceverilogrelpaths setexecmode
1
23,906
16,680,368,078
IssuesEvent
2021-06-07 22:26:54
google/site-kit-wp
https://api.github.com/repos/google/site-kit-wp
opened
getSetupFlowMode requires other selectors to be run
P2 QA: Eng Type: Infrastructure
## Bug Description The [`getSetupFlow`](https://github.com/google/site-kit-wp/blob/b252720322e3dce7883c92c79811c6d83e19f20c/assets/js/modules/analytics/datastore/setup-flow.js#L36) selector in the Analytics datastore has an issue that requires components that use it to call another selector (to resolve) first, before using `getSetupFlowMode`. This does not behave as-expected—any selector should have, in its resolvers or as a `createRegistrySelector`—all of its dependencies called so it can be used in a component without calling its dependant selectors first. ## Additional Context See discussion here: --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * `getSetupFlowMode` should be able to be called without first calling `select( MODULES_ANALYTICS_4 ).getProperties( accountID );` ## Implementation Brief * Remove https://github.com/google/site-kit-wp/blob/b252720322e3dce7883c92c79811c6d83e19f20c/assets/js/modules/analytics/components/settings/SettingsForm.js#L58-L66 ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the issue is resolved. --> ### Visual Regression Changes * <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. --> ## QA Brief * <!-- One or more bullet points for how to test that the issue has been resolved. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
1.0
getSetupFlowMode requires other selectors to be run - ## Bug Description The [`getSetupFlow`](https://github.com/google/site-kit-wp/blob/b252720322e3dce7883c92c79811c6d83e19f20c/assets/js/modules/analytics/datastore/setup-flow.js#L36) selector in the Analytics datastore has an issue that requires components that use it to call another selector (to resolve) first, before using `getSetupFlowMode`. This does not behave as-expected—any selector should have, in its resolvers or as a `createRegistrySelector`—all of its dependencies called so it can be used in a component without calling its dependant selectors first. ## Additional Context See discussion here: --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * `getSetupFlowMode` should be able to be called without first calling `select( MODULES_ANALYTICS_4 ).getProperties( accountID );` ## Implementation Brief * Remove https://github.com/google/site-kit-wp/blob/b252720322e3dce7883c92c79811c6d83e19f20c/assets/js/modules/analytics/components/settings/SettingsForm.js#L58-L66 ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the issue is resolved. --> ### Visual Regression Changes * <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. --> ## QA Brief * <!-- One or more bullet points for how to test that the issue has been resolved. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
infrastructure
getsetupflowmode requires other selectors to be run bug description the selector in the analytics datastore has an issue that requires components that use it to call another selector to resolve first before using getsetupflowmode this does not behave as expected—any selector should have in its resolvers or as a createregistryselector —all of its dependencies called so it can be used in a component without calling its dependant selectors first additional context see discussion here do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria getsetupflowmode should be able to be called without first calling select modules analytics getproperties accountid implementation brief remove test coverage visual regression changes qa brief changelog entry
1
17,589
12,474,634,039
IssuesEvent
2020-05-29 10:00:33
wellcomecollection/platform
https://api.github.com/repos/wellcomecollection/platform
reopened
Set up an S3 inventory report for the storage service
📦 Storage service 🚧 Infrastructure
See: * https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-inventory.html * https://www.terraform.io/docs/providers/aws/r/s3_bucket_inventory.html * https://wellcome.slack.com/archives/CN56BRQ5B/p1590499831057100 If you ask CloudWatch Metrics, we have slightly more objects in hot storage than cold, which seems odd. Would cost a few dollars a month, take maybe ten minutes to set up.
1.0
Set up an S3 inventory report for the storage service - See: * https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-inventory.html * https://www.terraform.io/docs/providers/aws/r/s3_bucket_inventory.html * https://wellcome.slack.com/archives/CN56BRQ5B/p1590499831057100 If you ask CloudWatch Metrics, we have slightly more objects in hot storage than cold, which seems odd. Would cost a few dollars a month, take maybe ten minutes to set up.
infrastructure
set up an inventory report for the storage service see if you ask cloudwatch metrics we have slightly more objects in hot storage than cold which seems odd would cost a few dollars a month take maybe ten minutes to set up
1
102,645
4,157,725,259
IssuesEvent
2016-06-16 22:22:40
Taeir/ContextProject-MIGI2
https://api.github.com/repos/Taeir/ContextProject-MIGI2
closed
Dynamically place lights (Task from last week) - Add dynamic lighting entity creation to corridors
Priority C
Corridors should contain light sources spread around, so that the corridors are not dark. This is a subtask of #289 - Dynamically place lights (Task from last week)
1.0
Dynamically place lights (Task from last week) - Add dynamic lighting entity creation to corridors - Corridors should contain light sources spread around, so that the corridors are not dark. This is a subtask of #289 - Dynamically place lights (Task from last week)
non_infrastructure
dynamically place lights task from last week add dynamic lighting entity creation to corridors corridors should contain light sources spread around so that the corridors are not dark this is a subtask of dynamically place lights task from last week
0
12,800
9,958,687,156
IssuesEvent
2019-07-05 22:35:30
zooniverse/caesar
https://api.github.com/repos/zooniverse/caesar
opened
prefetch relevant reductions
enhancement infrastructure performance
Currently we prefetch all the reductions we might need for a running reducer, but then on a per-reducer basis fetch the relevant reductions (user reductions for a subject reduction or subject reductions for a user reduction) that were added for TESS. These should also be prefetched to help avoid deadlocks.
1.0
prefetch relevant reductions - Currently we prefetch all the reductions we might need for a running reducer, but then on a per-reducer basis fetch the relevant reductions (user reductions for a subject reduction or subject reductions for a user reduction) that were added for TESS. These should also be prefetched to help avoid deadlocks.
infrastructure
prefetch relevant reductions currently we prefetch all the reductions we might need for a running reducer but then on a per reducer basis fetch the relevant reductions user reductions for a subject reduction or subject reductions for a user reduction that were added for tess these should also be prefetched to help avoid deadlocks
1
141,689
18,989,486,551
IssuesEvent
2021-11-22 04:25:13
ChoeMinji/kubernetes-1.12.0
https://api.github.com/repos/ChoeMinji/kubernetes-1.12.0
opened
CVE-2018-17847 (High) detected in multiple libraries
security vulnerability
## CVE-2018-17847 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b>, <b>golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b>, <b>github.com/miekg/dns-v1.0.0</b></p></summary> <p> <details><summary><b>github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b></p></summary> <p>[mirror] Go supplementary network libraries</p> <p> Dependency Hierarchy: - :x: **github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e** (Vulnerable Library) </details> <details><summary><b>golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b></p></summary> <p>[mirror] Go supplementary network libraries</p> <p> Dependency Hierarchy: - :x: **golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e** (Vulnerable Library) </details> <details><summary><b>github.com/miekg/dns-v1.0.0</b></p></summary> <p>DNS library in Go</p> <p> Dependency Hierarchy: - :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/kubernetes-1.12.0/commit/cd87321da747fd23c6285641b10b0432943dfed6">cd87321da747fd23c6285641b10b0432943dfed6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The html package (aka x/net/html) through 2018-09-25 in Go mishandles <svg><template><desc><t><svg></template>, leading to a "panic: runtime error" (index out of range) in (*nodeStack).pop in node.go, called from (*parser).clearActiveFormattingElements, during an html.Parse call. <p>Publish Date: 2018-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-17847>CVE-2018-17847</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-17847 (High) detected in multiple libraries - ## CVE-2018-17847 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b>, <b>golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b>, <b>github.com/miekg/dns-v1.0.0</b></p></summary> <p> <details><summary><b>github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b></p></summary> <p>[mirror] Go supplementary network libraries</p> <p> Dependency Hierarchy: - :x: **github.com/golang/net-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e** (Vulnerable Library) </details> <details><summary><b>golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e</b></p></summary> <p>[mirror] Go supplementary network libraries</p> <p> Dependency Hierarchy: - :x: **golang.org/x/net/html-b129b8e0fbeb39c8358e51a07ab6c50ad415e72e** (Vulnerable Library) </details> <details><summary><b>github.com/miekg/dns-v1.0.0</b></p></summary> <p>DNS library in Go</p> <p> Dependency Hierarchy: - :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/kubernetes-1.12.0/commit/cd87321da747fd23c6285641b10b0432943dfed6">cd87321da747fd23c6285641b10b0432943dfed6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The html package (aka x/net/html) through 2018-09-25 in Go mishandles <svg><template><desc><t><svg></template>, leading to a "panic: runtime error" (index out of range) in (*nodeStack).pop in node.go, called from (*parser).clearActiveFormattingElements, during an html.Parse call. <p>Publish Date: 2018-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-17847>CVE-2018-17847</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries github com golang net golang org x net html github com miekg dns github com golang net go supplementary network libraries dependency hierarchy x github com golang net vulnerable library golang org x net html go supplementary network libraries dependency hierarchy x golang org x net html vulnerable library github com miekg dns dns library in go dependency hierarchy x github com miekg dns vulnerable library found in head commit a href found in base branch master vulnerability details the html package aka x net html through in go mishandles leading to a panic runtime error index out of range in nodestack pop in node go called from parser clearactiveformattingelements during an html parse call publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
208,456
15,890,853,326
IssuesEvent
2021-04-10 16:58:03
NickleDave/hybrid-vocal-classifier
https://api.github.com/repos/NickleDave/hybrid-vocal-classifier
closed
refactor test directory structure: rename test_data, put set-up scripts in a scripts dir
TST: testing
- [ ] rename `test_data` to `data_for_tests` + so it looks different from the directories of tests - [ ] put scripts that generate test data + `tests/utils` script in their own `scripts` directory within tests + so we can remove nasty path surgery at top of scripts that imports `config` module from `utils` + and so the scripts are not mixed in with the test data and we can just `.gitignore` the whole `data_for_tests` dir
1.0
refactor test directory structure: rename test_data, put set-up scripts in a scripts dir - - [ ] rename `test_data` to `data_for_tests` + so it looks different from the directories of tests - [ ] put scripts that generate test data + `tests/utils` script in their own `scripts` directory within tests + so we can remove nasty path surgery at top of scripts that imports `config` module from `utils` + and so the scripts are not mixed in with the test data and we can just `.gitignore` the whole `data_for_tests` dir
non_infrastructure
refactor test directory structure rename test data put set up scripts in a scripts dir rename test data to data for tests so it looks different from the directories of tests put scripts that generate test data tests utils script in their own scripts directory within tests so we can remove nasty path surgery at top of scripts that imports config module from utils and so the scripts are not mixed in with the test data and we can just gitignore the whole data for tests dir
0
356,980
25,176,300,716
IssuesEvent
2022-11-11 09:33:45
zsiggg/pe
https://api.github.com/repos/zsiggg/pe
opened
Repetitive sequence diagrams for add, edit, delete events
type.DocumentationBug severity.VeryLow
![Screenshot 2022-11-11 at 5.30.44 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/0a8f4d39-8d0a-401c-bfb5-292f42b69d23.png) ![Screenshot 2022-11-11 at 5.30.55 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/70780c29-467d-4464-9568-989970008ec6.png) ![Screenshot 2022-11-11 at 5.31.02 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/9c917f99-dedd-4c94-a4f4-3af4681e273d.png) The 3 sequence diagrams seem a bit repetitive and thus unnecessary, since they are all showing the same activation bars and arrows, except with a name change. Perhaps 1 of them would have been enough, with the other two diagram replaced by telling the reader that it has a similar flow to that diagram. <!--session: 1668152772162-768faedc-ab16-41ca-9507-bd1a73b2c58e--> <!--Version: Web v3.4.4-->
1.0
Repetitive sequence diagrams for add, edit, delete events - ![Screenshot 2022-11-11 at 5.30.44 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/0a8f4d39-8d0a-401c-bfb5-292f42b69d23.png) ![Screenshot 2022-11-11 at 5.30.55 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/70780c29-467d-4464-9568-989970008ec6.png) ![Screenshot 2022-11-11 at 5.31.02 PM.png](https://raw.githubusercontent.com/zsiggg/pe/main/files/9c917f99-dedd-4c94-a4f4-3af4681e273d.png) The 3 sequence diagrams seem a bit repetitive and thus unnecessary, since they are all showing the same activation bars and arrows, except with a name change. Perhaps 1 of them would have been enough, with the other two diagram replaced by telling the reader that it has a similar flow to that diagram. <!--session: 1668152772162-768faedc-ab16-41ca-9507-bd1a73b2c58e--> <!--Version: Web v3.4.4-->
non_infrastructure
repetitive sequence diagrams for add edit delete events the sequence diagrams seem a bit repetitive and thus unnecessary since they are all showing the same activation bars and arrows except with a name change perhaps of them would have been enough with the other two diagram replaced by telling the reader that it has a similar flow to that diagram
0
123,871
17,772,361,766
IssuesEvent
2021-08-30 15:00:28
kapseliboi/spotlight
https://api.github.com/repos/kapseliboi/spotlight
opened
CVE-2020-7662 (High) detected in websocket-extensions-0.1.3.tgz
security vulnerability
## CVE-2020-7662 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websocket-extensions-0.1.3.tgz</b></p></summary> <p>Generic extension manager for WebSocket connections</p> <p>Library home page: <a href="https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz">https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz</a></p> <p>Path to dependency file: spotlight/package.json</p> <p>Path to vulnerable library: spotlight/node_modules/websocket-extensions/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-1.0.0.tgz (Root Library) - tiny-lr-0.2.1.tgz - faye-websocket-0.10.0.tgz - websocket-driver-0.7.0.tgz - :x: **websocket-extensions-0.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/spotlight/commit/efa8ebd4395408150b8ea1a18eec77751d13827b">efa8ebd4395408150b8ea1a18eec77751d13827b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> websocket-extensions npm module prior to 0.1.4 allows Denial of Service (DoS) via Regex Backtracking. The extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two-byte sequence of a backslash and some other character. This could be abused by an attacker to conduct Regex Denial Of Service (ReDoS) on a single-threaded server by providing a malicious payload with the Sec-WebSocket-Extensions header. <p>Publish Date: 2020-06-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662>CVE-2020-7662</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-g78m-2chm-r7qv">https://github.com/advisories/GHSA-g78m-2chm-r7qv</a></p> <p>Release Date: 2020-06-02</p> <p>Fix Resolution: websocket-extensions - 0.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7662 (High) detected in websocket-extensions-0.1.3.tgz - ## CVE-2020-7662 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websocket-extensions-0.1.3.tgz</b></p></summary> <p>Generic extension manager for WebSocket connections</p> <p>Library home page: <a href="https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz">https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz</a></p> <p>Path to dependency file: spotlight/package.json</p> <p>Path to vulnerable library: spotlight/node_modules/websocket-extensions/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-1.0.0.tgz (Root Library) - tiny-lr-0.2.1.tgz - faye-websocket-0.10.0.tgz - websocket-driver-0.7.0.tgz - :x: **websocket-extensions-0.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/spotlight/commit/efa8ebd4395408150b8ea1a18eec77751d13827b">efa8ebd4395408150b8ea1a18eec77751d13827b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> websocket-extensions npm module prior to 0.1.4 allows Denial of Service (DoS) via Regex Backtracking. The extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two-byte sequence of a backslash and some other character. This could be abused by an attacker to conduct Regex Denial Of Service (ReDoS) on a single-threaded server by providing a malicious payload with the Sec-WebSocket-Extensions header. <p>Publish Date: 2020-06-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662>CVE-2020-7662</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-g78m-2chm-r7qv">https://github.com/advisories/GHSA-g78m-2chm-r7qv</a></p> <p>Release Date: 2020-06-02</p> <p>Fix Resolution: websocket-extensions - 0.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in websocket extensions tgz cve high severity vulnerability vulnerable library websocket extensions tgz generic extension manager for websocket connections library home page a href path to dependency file spotlight package json path to vulnerable library spotlight node modules websocket extensions package json dependency hierarchy grunt contrib watch tgz root library tiny lr tgz faye websocket tgz websocket driver tgz x websocket extensions tgz vulnerable library found in head commit a href found in base branch master vulnerability details websocket extensions npm module prior to allows denial of service dos via regex backtracking the extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two byte sequence of a backslash and some other character this could be abused by an attacker to conduct regex denial of service redos on a single threaded server by providing a malicious payload with the sec websocket extensions header publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution websocket extensions step up your open source security game with whitesource
0
656
9,540,304,183
IssuesEvent
2019-04-30 19:10:23
Microsoft/VFSForGit
https://api.github.com/repos/Microsoft/VFSForGit
closed
Uninstalling VFS For Git doesn't clean up filter driver
affects: reliability platform: windows type: bug
Running the VFS For Git uninstaller doesn't cleanup the projfs filter driver and screws up the following VFS For Git installation. We've hit this issue while trying to upgrade several Azure agent pool machines.
True
Uninstalling VFS For Git doesn't clean up filter driver - Running the VFS For Git uninstaller doesn't cleanup the projfs filter driver and screws up the following VFS For Git installation. We've hit this issue while trying to upgrade several Azure agent pool machines.
non_infrastructure
uninstalling vfs for git doesn t clean up filter driver running the vfs for git uninstaller doesn t cleanup the projfs filter driver and screws up the following vfs for git installation we ve hit this issue while trying to upgrade several azure agent pool machines
0
74,894
9,811,318,740
IssuesEvent
2019-06-12 23:13:38
brusMX/promitor-tests
https://api.github.com/repos/brusMX/promitor-tests
opened
Follow README
documentation wip
Make sure README makes sense, documentation is on par, I can do the listed tasks
1.0
Follow README - Make sure README makes sense, documentation is on par, I can do the listed tasks
non_infrastructure
follow readme make sure readme makes sense documentation is on par i can do the listed tasks
0
15,603
11,610,247,107
IssuesEvent
2020-02-26 02:23:21
algorand/go-algorand
https://api.github.com/repos/algorand/go-algorand
closed
MainNet Telemetry Error: pseudonode.makeProposals: could not generate a proposal ... ledger: ledger does not have entry
1 infrastructure
(originally reported by @Karmastic July 15 but this was seen as late as Aug 7) 7 in last week; 6 from same node, 1 from r1. In an effort to reduce telemetry errors, we should investigate and see if we can re-categorize this. `pseudonode.makeProposals: could not generate a proposal for round 607841: could not make proposals at round 607841: could not read block from ledger: ledger does not have entry`
1.0
MainNet Telemetry Error: pseudonode.makeProposals: could not generate a proposal ... ledger: ledger does not have entry - (originally reported by @Karmastic July 15 but this was seen as late as Aug 7) 7 in last week; 6 from same node, 1 from r1. In an effort to reduce telemetry errors, we should investigate and see if we can re-categorize this. `pseudonode.makeProposals: could not generate a proposal for round 607841: could not make proposals at round 607841: could not read block from ledger: ledger does not have entry`
infrastructure
mainnet telemetry error pseudonode makeproposals could not generate a proposal ledger ledger does not have entry originally reported by karmastic july but this was seen as late as aug in last week from same node from in an effort to reduce telemetry errors we should investigate and see if we can re categorize this pseudonode makeproposals could not generate a proposal for round could not make proposals at round could not read block from ledger ledger does not have entry
1
21,348
14,532,532,480
IssuesEvent
2020-12-14 22:36:17
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
reopened
Expand 'edit apsimx file' functionality
interface/infrastructure newfeature
Need to be able to replace a model from an external file via the command line when using the /Edit command-line switch.
1.0
Expand 'edit apsimx file' functionality - Need to be able to replace a model from an external file via the command line when using the /Edit command-line switch.
infrastructure
expand edit apsimx file functionality need to be able to replace a model from an external file via the command line when using the edit command line switch
1
90,838
15,856,286,848
IssuesEvent
2021-04-08 01:59:11
ChenLuigi/Test_Npm__CircleCi
https://api.github.com/repos/ChenLuigi/Test_Npm__CircleCi
opened
CVE-2020-8244 (Medium) detected in bl-1.2.2.tgz
security vulnerability
## CVE-2020-8244 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bl-1.2.2.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.2.tgz">https://registry.npmjs.org/bl/-/bl-1.2.2.tgz</a></p> <p>Path to dependency file: /Test_Npm__CircleCi/package.json</p> <p>Path to vulnerable library: Test_Npm__CircleCi/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - download-6.2.5.tgz (Root Library) - decompress-4.2.0.tgz - decompress-tar-4.1.1.tgz - tar-stream-1.6.1.tgz - :x: **bl-1.2.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls. <p>Publish Date: 2020-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.2.1,3.0.1,4.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8244 (Medium) detected in bl-1.2.2.tgz - ## CVE-2020-8244 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bl-1.2.2.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.2.tgz">https://registry.npmjs.org/bl/-/bl-1.2.2.tgz</a></p> <p>Path to dependency file: /Test_Npm__CircleCi/package.json</p> <p>Path to vulnerable library: Test_Npm__CircleCi/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - download-6.2.5.tgz (Root Library) - decompress-4.2.0.tgz - decompress-tar-4.1.1.tgz - tar-stream-1.6.1.tgz - :x: **bl-1.2.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls. <p>Publish Date: 2020-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8244</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.2.1,3.0.1,4.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in bl tgz cve medium severity vulnerability vulnerable library bl tgz buffer list collect buffers and access with a standard readable buffer interface streamable too library home page a href path to dependency file test npm circleci package json path to vulnerable library test npm circleci node modules bl package json dependency hierarchy download tgz root library decompress tgz decompress tar tgz tar stream tgz x bl tgz vulnerable library vulnerability details a buffer over read vulnerability exists in bl and which could allow an attacker to supply user input even typed that if it ends up in consume argument and can become negative the bufferlist state can be corrupted tricking it into exposing uninitialized memory via regular slice calls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
106,094
9,107,750,730
IssuesEvent
2019-02-21 06:00:54
xcat2/xcat2-task-management
https://api.github.com/repos/xcat2/xcat2-task-management
closed
test case makehost_n_r need to refine
sprint3 test
Test case ``makehost_n_r`` failed frequently in ubuntu. Due to xcatd leverage cache mechanism, so ``makehost`` maybe fail to read data from cache. that results in makehost command failed. Need to refine test case `` makehost_n_r `` to cover this situation. ``` ------START::makehost_n_r::Time:Mon Feb 18 21:39:36 2019------ FILENAME:/opt/xcat/bin/../share/xcat/tools/autotest/testcase//makehosts/cases0 RUN:cp -f /etc/hosts /etc/hosts.xcatbakautotest [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ ..... 100.100.100.2 testnode RUN:lsdef s01;if [ $? -eq 0 ]; then lsdef -l s01 -z >/tmp/s01.standa ;rmdef s01;fi [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: No object definitions have been found CHECK:rc == 0 [Pass] RUN:lsdef s01r1b01;if [ $? -eq 0 ]; then lsdef -l s01r1b01 -z >/tmp/s01r1b01.standa ;rmdef s01r1b01;fi [Mon Feb 18 21:39:36 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: No object definitions have been found CHECK:rc == 0 [Pass] RUN:nodeadd s01 groups=service; chdef s01 ip=70.2.0.254;nodeadd s01r1b01 groups=compute; chdef s01r1b01 ip=80.2.0.254 [Mon Feb 18 21:39:37 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: 1 object definitions have been created or modified. 1 object definitions have been created or modified. CHECK:rc == 0 [Pass] RUN:lsdef -l s01,s01r1b01 [Mon Feb 18 21:39:38 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: Object name: s01 groups=service ip=70.2.0.254 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles,servicenode Object name: s01r1b01 groups=compute ip=80.2.0.254 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles RUN:makehosts [Mon Feb 18 21:39:39 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: CHECK:rc == 0 [Pass] RUN:cat /etc/hosts [Mon Feb 18 21:39:39 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ .... 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:39:39 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "Do not find s01 and s01r1b01 in $file, sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:30 sec RETURN rc = 1 OUTPUT: Do not find s01 and s01r1b01 in /etc/hosts, sleep 2 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 4 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 6 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 8 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 10 seconds and try again CHECK:rc == 0 [Failed] RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts s01 [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:09 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && ( ! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts service [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:09 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts -d s01 [Mon Feb 18 21:40:09 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ .... 100.100.100.2 testnode RUN: [Mon Feb 18 21:40:10 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (! grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN:makehosts -d service [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ .... 10.3.17.7 ports.ubuntu.com 10.3.17.7 archive.ubuntu.com 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:10 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (! grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts [Mon Feb 18 21:40:10 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN:makehosts -d s01r1b01 [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system ...... 0.3.17.7 ports.ubuntu.com 10.3.17.7 archive.ubuntu.com 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:11 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:if [ -e /tmp/s01.standa ]; then rmdef s01; cat /tmp/s01.standa | mkdef -z; rm -rf /tmp/s01.standa; else rmdef s01;fi [Mon Feb 18 21:40:11 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: 1 object definitions have been removed. RUN:if [ -e /tmp/s01r1b01.standa ]; then rmdef s01r1b01; cat /tmp/s01r1b01.standa | mkdef -z; rm -rf /tmp/s01r1b01.standa;else rmdef s01r1b01; fi [Mon Feb 18 21:40:12 2019] ElapsedTime:3 sec RETURN rc = 0 OUTPUT: 1 object definitions have been removed. RUN:mv -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:15 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: ------END::makehost_n_r::Failed::Time:Mon Feb 18 21:40:15 2019 ::Duration::39 sec------ ```
1.0
test case makehost_n_r need to refine - Test case ``makehost_n_r`` failed frequently in ubuntu. Due to xcatd leverage cache mechanism, so ``makehost`` maybe fail to read data from cache. that results in makehost command failed. Need to refine test case `` makehost_n_r `` to cover this situation. ``` ------START::makehost_n_r::Time:Mon Feb 18 21:39:36 2019------ FILENAME:/opt/xcat/bin/../share/xcat/tools/autotest/testcase//makehosts/cases0 RUN:cp -f /etc/hosts /etc/hosts.xcatbakautotest [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ ..... 100.100.100.2 testnode RUN:lsdef s01;if [ $? -eq 0 ]; then lsdef -l s01 -z >/tmp/s01.standa ;rmdef s01;fi [Mon Feb 18 21:39:36 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: No object definitions have been found CHECK:rc == 0 [Pass] RUN:lsdef s01r1b01;if [ $? -eq 0 ]; then lsdef -l s01r1b01 -z >/tmp/s01r1b01.standa ;rmdef s01r1b01;fi [Mon Feb 18 21:39:36 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: No object definitions have been found CHECK:rc == 0 [Pass] RUN:nodeadd s01 groups=service; chdef s01 ip=70.2.0.254;nodeadd s01r1b01 groups=compute; chdef s01r1b01 ip=80.2.0.254 [Mon Feb 18 21:39:37 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: 1 object definitions have been created or modified. 1 object definitions have been created or modified. CHECK:rc == 0 [Pass] RUN:lsdef -l s01,s01r1b01 [Mon Feb 18 21:39:38 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: Object name: s01 groups=service ip=70.2.0.254 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles,servicenode Object name: s01r1b01 groups=compute ip=80.2.0.254 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles RUN:makehosts [Mon Feb 18 21:39:39 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: CHECK:rc == 0 [Pass] RUN:cat /etc/hosts [Mon Feb 18 21:39:39 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ .... 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:39:39 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "Do not find s01 and s01r1b01 in $file, sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:30 sec RETURN rc = 1 OUTPUT: Do not find s01 and s01r1b01 in /etc/hosts, sleep 2 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 4 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 6 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 8 seconds and try again Do not find s01 and s01r1b01 in /etc/hosts, sleep 10 seconds and try again CHECK:rc == 0 [Failed] RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts s01 [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:09 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && ( ! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts service [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:09 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:09 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts -d s01 [Mon Feb 18 21:40:09 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ .... 100.100.100.2 testnode RUN: [Mon Feb 18 21:40:10 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (! grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 70.2.0.254 s01 s01.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN:makehosts -d service [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system # network 10.0.0.0 # netmask 255.0.0.0 # gateway 10.255.255.254 ############################################################ .... 10.3.17.7 ports.ubuntu.com 10.3.17.7 archive.ubuntu.com 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:10 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (! grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cp -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:10 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:makehosts [Mon Feb 18 21:40:10 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ ..... 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com 80.2.0.254 s01r1b01 s01r1b01.pok.stglabs.ibm.com RUN:makehosts -d s01r1b01 [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:cat /etc/hosts [Mon Feb 18 21:40:11 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 ############################################################ # c910 xCAT Private network system ...... 0.3.17.7 ports.ubuntu.com 10.3.17.7 archive.ubuntu.com 100.100.100.2 testnode 10.3.17.39 c910f03c17k39 c910f03c17k39.pok.stglabs.ibm.com 70.2.0.254 s01 s01.pok.stglabs.ibm.com RUN: [Mon Feb 18 21:40:11 2019] #!/bin/bash file="/etc/hosts" for i in {1..5}; do if (grep "70.2.0.254 s01" $file >/dev/null 2>&1) && (! grep "80.2.0.254 s01r1b01" $file >/dev/null 2>&1); then exit 0; else echo "sleep $[i*2] seconds and try again" sleep $[i*2] fi done exit 1 ElapsedTime:0 sec RETURN rc = 0 OUTPUT: RUN:if [ -e /tmp/s01.standa ]; then rmdef s01; cat /tmp/s01.standa | mkdef -z; rm -rf /tmp/s01.standa; else rmdef s01;fi [Mon Feb 18 21:40:11 2019] ElapsedTime:1 sec RETURN rc = 0 OUTPUT: 1 object definitions have been removed. RUN:if [ -e /tmp/s01r1b01.standa ]; then rmdef s01r1b01; cat /tmp/s01r1b01.standa | mkdef -z; rm -rf /tmp/s01r1b01.standa;else rmdef s01r1b01; fi [Mon Feb 18 21:40:12 2019] ElapsedTime:3 sec RETURN rc = 0 OUTPUT: 1 object definitions have been removed. RUN:mv -f /etc/hosts.xcatbakautotest /etc/hosts [Mon Feb 18 21:40:15 2019] ElapsedTime:0 sec RETURN rc = 0 OUTPUT: ------END::makehost_n_r::Failed::Time:Mon Feb 18 21:40:15 2019 ::Duration::39 sec------ ```
non_infrastructure
test case makehost n r need to refine test case makehost n r failed frequently in ubuntu due to xcatd leverage cache mechanism so makehost maybe fail to read data from cache that results in makehost command failed need to refine test case makehost n r to cover this situation start makehost n r time mon feb filename opt xcat bin share xcat tools autotest testcase makehosts run cp f etc hosts etc hosts xcatbakautotest elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain xcat private network system network netmask gateway testnode run lsdef if then lsdef l z tmp standa rmdef fi elapsedtime sec return rc output no object definitions have been found check rc run lsdef if then lsdef l z tmp standa rmdef fi elapsedtime sec return rc output no object definitions have been found check rc run nodeadd groups service chdef ip nodeadd groups compute chdef ip elapsedtime sec return rc output object definitions have been created or modified object definitions have been created or modified check rc run lsdef l elapsedtime sec return rc output object name groups service ip postbootscripts otherpkgs postscripts syslog remoteshell syncfiles servicenode object name groups compute ip postbootscripts otherpkgs postscripts syslog remoteshell syncfiles run makehosts elapsedtime sec return rc output check rc run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain xcat private network system network netmask gateway testnode pok stglabs ibm com pok stglabs ibm com run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo do not find and in file sleep seconds and try again sleep fi done exit elapsedtime sec return rc output do not find and in etc hosts sleep seconds and try again do not find and in etc hosts sleep seconds and try again do not find and in etc hosts sleep seconds and try again do not find and in etc hosts sleep seconds and try again do not find and in etc hosts sleep seconds and try again check rc run cp f etc hosts xcatbakautotest etc hosts elapsedtime sec return rc output run makehosts elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain pok stglabs ibm com run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo sleep seconds and try again sleep fi done exit elapsedtime sec return rc output run cp f etc hosts xcatbakautotest etc hosts elapsedtime sec return rc output run makehosts service elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain pok stglabs ibm com run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo sleep seconds and try again sleep fi done exit elapsedtime sec return rc output run makehosts d elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain testnode run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo sleep seconds and try again sleep fi done exit elapsedtime sec return rc output run cp f etc hosts xcatbakautotest etc hosts elapsedtime sec return rc output run makehosts elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain pok stglabs ibm com pok stglabs ibm com run makehosts d service elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain xcat private network system network netmask gateway ports ubuntu com archive ubuntu com testnode pok stglabs ibm com pok stglabs ibm com run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo sleep seconds and try again sleep fi done exit elapsedtime sec return rc output run cp f etc hosts xcatbakautotest etc hosts elapsedtime sec return rc output run makehosts elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain testnode pok stglabs ibm com pok stglabs ibm com pok stglabs ibm com run makehosts d elapsedtime sec return rc output run cat etc hosts elapsedtime sec return rc output localhost localhost localdomain localhost localhost localdomain xcat private network system ports ubuntu com archive ubuntu com testnode pok stglabs ibm com pok stglabs ibm com run bin bash file etc hosts for i in do if grep file dev null grep file dev null then exit else echo sleep seconds and try again sleep fi done exit elapsedtime sec return rc output run if then rmdef cat tmp standa mkdef z rm rf tmp standa else rmdef fi elapsedtime sec return rc output object definitions have been removed run if then rmdef cat tmp standa mkdef z rm rf tmp standa else rmdef fi elapsedtime sec return rc output object definitions have been removed run mv f etc hosts xcatbakautotest etc hosts elapsedtime sec return rc output end makehost n r failed time mon feb duration sec
0
4,831
5,291,299,336
IssuesEvent
2017-02-08 22:08:51
mirai-audio/mir
https://api.github.com/repos/mirai-audio/mir
reopened
Setup contact endpoints
infrastructure
# Goal / Issue Allow users, collaborators and others to contact the project and its maintainers. ## Tasks * [x] Find a free mail hosting provider * [x] Setup mailbox for new email * [x] Setup mail forwarding to project owners * [x] Configure DNS `mx` records on Cloudflare to point to mail provider * [x] Create twitter account * [x] Enable 2FA
1.0
Setup contact endpoints - # Goal / Issue Allow users, collaborators and others to contact the project and its maintainers. ## Tasks * [x] Find a free mail hosting provider * [x] Setup mailbox for new email * [x] Setup mail forwarding to project owners * [x] Configure DNS `mx` records on Cloudflare to point to mail provider * [x] Create twitter account * [x] Enable 2FA
infrastructure
setup contact endpoints goal issue allow users collaborators and others to contact the project and its maintainers tasks find a free mail hosting provider setup mailbox for new email setup mail forwarding to project owners configure dns mx records on cloudflare to point to mail provider create twitter account enable
1
27,585
21,943,385,650
IssuesEvent
2022-05-23 20:42:09
Shopify/polaris
https://api.github.com/repos/Shopify/polaris
closed
Simplify release and changelog process
🗺 Roadmap Infrastructure
The [Polaris release process](https://docs.google.com/document/d/1sjbYdZb1VqhZori-WlmqrpiZW-Xx7xLBbFekbvFIU4g/edit) is outdated, hard to follow and often causes issues with tagging releases and shipping to NPM. It has slowed down multiple product teams (Deliver) from efficiently shipping new features to shopify/web. Getting the release process down to 2 steps will increase velocity, reduce forked components, increase contributions and allow our team to work with product teams more efficiently. - [x] Remove release branch - [x] Remove [changelog.md](https://github.com/Shopify/polaris-react/blob/main/CHANGELOG.md) and [unreleased.md](https://github.com/Shopify/polaris-react/blob/main/UNRELEASED.md) use [auto-generate release notes](https://docs.github.com/en/repositories/releasing-projects-on-github/automatically-generated-release-notes) - [x] Tag commits through the release UI - [x] When release published from GitHub UI automation is ran to: 1. bump package.json 2. publish NPM package (either with button press on shipit or automation) - [x] Document release process in README.md file (alpha, patch, minor, major) - [x] Communicate release process to teams
1.0
Simplify release and changelog process - The [Polaris release process](https://docs.google.com/document/d/1sjbYdZb1VqhZori-WlmqrpiZW-Xx7xLBbFekbvFIU4g/edit) is outdated, hard to follow and often causes issues with tagging releases and shipping to NPM. It has slowed down multiple product teams (Deliver) from efficiently shipping new features to shopify/web. Getting the release process down to 2 steps will increase velocity, reduce forked components, increase contributions and allow our team to work with product teams more efficiently. - [x] Remove release branch - [x] Remove [changelog.md](https://github.com/Shopify/polaris-react/blob/main/CHANGELOG.md) and [unreleased.md](https://github.com/Shopify/polaris-react/blob/main/UNRELEASED.md) use [auto-generate release notes](https://docs.github.com/en/repositories/releasing-projects-on-github/automatically-generated-release-notes) - [x] Tag commits through the release UI - [x] When release published from GitHub UI automation is ran to: 1. bump package.json 2. publish NPM package (either with button press on shipit or automation) - [x] Document release process in README.md file (alpha, patch, minor, major) - [x] Communicate release process to teams
infrastructure
simplify release and changelog process the is outdated hard to follow and often causes issues with tagging releases and shipping to npm it has slowed down multiple product teams deliver from efficiently shipping new features to shopify web getting the release process down to steps will increase velocity reduce forked components increase contributions and allow our team to work with product teams more efficiently remove release branch remove and use tag commits through the release ui when release published from github ui automation is ran to bump package json publish npm package either with button press on shipit or automation document release process in readme md file alpha patch minor major communicate release process to teams
1
31,498
25,818,723,117
IssuesEvent
2022-12-12 07:54:18
Stellarium/stellarium
https://api.github.com/repos/Stellarium/stellarium
closed
Cannot install v1.1 on Mac OS 10.5.7
infrastructure state: published
<!-- READ THIS AND FILL IN THIS TEMPLATE! Before reporting an issue around graphics artifacts like missing menu buttons, strange colors, Moon not rendered, or similar, *please make absolutely sure* you are running the latest graphics drivers for your graphics card. 95% of graphics problems are solved this way. Before reporting an issue, please *absolutely make sure* to check the recent open and also closed issues (https://github.com/Stellarium/stellarium/issues?q=is%3Aissue+is%3Aclosed) whether it has been reported and solved/closed already! Don't report a new issue in this case, you may join the discussion. Sometimes installing our almost weekly Beta version (i.e., the latest development version) helps: https://github.com/Stellarium/stellarium-data/releases/tag/beta Also look into the wiki https://github.com/Stellarium/stellarium/wiki/Common-Problems-for-the-current-version and FAQ https://github.com/Stellarium/stellarium/wiki/FAQ and check if you are attempting to report a known issue. Don't report in this case. Also look into the User Guide before reporting unexpected but correct behaviour. Don't report in this case. If you miss translations, please help us with your language and join our translators at Transifex: https://www.transifex.com/stellarium/stellarium/dashboard/ If you want to report a feature wish, delete these template lines above and below, and fill in. If you want to report a bug, please make sure to update and verify the bug still exists in the current version. In this case please also check current beta builds from http://stellarium.org If the bug still persists, give your issue a clear, short title, delete text above down to and including this line and fill in: --> ### Expected Behaviour Able to be installed. 10.15.7 was last updated in July 2022 and is the latest version I can install. If there is nothing that can be done about this then how do I reinstall the previous version? ### Actual Behaviour Complains about not being compatible. ### Steps to reproduce Attempt to install it. ### System * Stellarium version: <v1.1> * Operating system: <Mac OS, 10.15.7> * Graphics Card: <Manufacturer (likely Intel, NVidia, AMD?), Model (HD, Geforce, Radeon..., with model number), driver version?> * Screen type (if applicable): Resolution, HighDPI, scaling ### Logfile If possible, attach the logfile `log.txt` from your user data directory. Look into the Guide for its location.
1.0
Cannot install v1.1 on Mac OS 10.5.7 - <!-- READ THIS AND FILL IN THIS TEMPLATE! Before reporting an issue around graphics artifacts like missing menu buttons, strange colors, Moon not rendered, or similar, *please make absolutely sure* you are running the latest graphics drivers for your graphics card. 95% of graphics problems are solved this way. Before reporting an issue, please *absolutely make sure* to check the recent open and also closed issues (https://github.com/Stellarium/stellarium/issues?q=is%3Aissue+is%3Aclosed) whether it has been reported and solved/closed already! Don't report a new issue in this case, you may join the discussion. Sometimes installing our almost weekly Beta version (i.e., the latest development version) helps: https://github.com/Stellarium/stellarium-data/releases/tag/beta Also look into the wiki https://github.com/Stellarium/stellarium/wiki/Common-Problems-for-the-current-version and FAQ https://github.com/Stellarium/stellarium/wiki/FAQ and check if you are attempting to report a known issue. Don't report in this case. Also look into the User Guide before reporting unexpected but correct behaviour. Don't report in this case. If you miss translations, please help us with your language and join our translators at Transifex: https://www.transifex.com/stellarium/stellarium/dashboard/ If you want to report a feature wish, delete these template lines above and below, and fill in. If you want to report a bug, please make sure to update and verify the bug still exists in the current version. In this case please also check current beta builds from http://stellarium.org If the bug still persists, give your issue a clear, short title, delete text above down to and including this line and fill in: --> ### Expected Behaviour Able to be installed. 10.15.7 was last updated in July 2022 and is the latest version I can install. If there is nothing that can be done about this then how do I reinstall the previous version? ### Actual Behaviour Complains about not being compatible. ### Steps to reproduce Attempt to install it. ### System * Stellarium version: <v1.1> * Operating system: <Mac OS, 10.15.7> * Graphics Card: <Manufacturer (likely Intel, NVidia, AMD?), Model (HD, Geforce, Radeon..., with model number), driver version?> * Screen type (if applicable): Resolution, HighDPI, scaling ### Logfile If possible, attach the logfile `log.txt` from your user data directory. Look into the Guide for its location.
infrastructure
cannot install on mac os read this and fill in this template before reporting an issue around graphics artifacts like missing menu buttons strange colors moon not rendered or similar please make absolutely sure you are running the latest graphics drivers for your graphics card of graphics problems are solved this way before reporting an issue please absolutely make sure to check the recent open and also closed issues whether it has been reported and solved closed already don t report a new issue in this case you may join the discussion sometimes installing our almost weekly beta version i e the latest development version helps also look into the wiki and faq and check if you are attempting to report a known issue don t report in this case also look into the user guide before reporting unexpected but correct behaviour don t report in this case if you miss translations please help us with your language and join our translators at transifex if you want to report a feature wish delete these template lines above and below and fill in if you want to report a bug please make sure to update and verify the bug still exists in the current version in this case please also check current beta builds from if the bug still persists give your issue a clear short title delete text above down to and including this line and fill in expected behaviour able to be installed was last updated in july and is the latest version i can install if there is nothing that can be done about this then how do i reinstall the previous version actual behaviour complains about not being compatible steps to reproduce attempt to install it system stellarium version operating system graphics card screen type if applicable resolution highdpi scaling logfile if possible attach the logfile log txt from your user data directory look into the guide for its location
1
2,798
3,872,870,501
IssuesEvent
2016-04-11 15:10:45
OpenSCAP/scap-security-guide
https://api.github.com/repos/OpenSCAP/scap-security-guide
closed
[BugFix] The Value 'oval:ssg-no line is returned:var:1' is not facet-valid with respect to pattern 'ocil:[A-Za-z0-9_\-\.]+:variable:[1-9][0-9]*' for type 'VariableIDPattern'
bugfix Infrastructure
First issue reported from Dragos Prisaca of NIST during performing SSG content ScapVal review: ``` The Value 'oval:ssg-no line is returned:var:1' is not facet-valid with respect to pattern 'ocil:[A-Za-z0-9_\-\.]+:variable:[1-9][0-9]*' for type 'VariableIDPattern' (please see the XCCDF 1.2 schema). ```
1.0
[BugFix] The Value 'oval:ssg-no line is returned:var:1' is not facet-valid with respect to pattern 'ocil:[A-Za-z0-9_\-\.]+:variable:[1-9][0-9]*' for type 'VariableIDPattern' - First issue reported from Dragos Prisaca of NIST during performing SSG content ScapVal review: ``` The Value 'oval:ssg-no line is returned:var:1' is not facet-valid with respect to pattern 'ocil:[A-Za-z0-9_\-\.]+:variable:[1-9][0-9]*' for type 'VariableIDPattern' (please see the XCCDF 1.2 schema). ```
infrastructure
the value oval ssg no line is returned var is not facet valid with respect to pattern ocil variable for type variableidpattern first issue reported from dragos prisaca of nist during performing ssg content scapval review the value oval ssg no line is returned var is not facet valid with respect to pattern ocil variable for type variableidpattern please see the xccdf schema
1
35,009
30,676,637,432
IssuesEvent
2023-07-26 06:01:43
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[arm32/arm64] Enable IL asm round-trip testing in CI
arch-arm32 arch-arm64 area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate
The following job needs to be enabled for arm/armlb/arm64: ``` 'ilrt' ``` This is a similar issue to dotnet/runtime#9348.
1.0
[arm32/arm64] Enable IL asm round-trip testing in CI - The following job needs to be enabled for arm/armlb/arm64: ``` 'ilrt' ``` This is a similar issue to dotnet/runtime#9348.
infrastructure
enable il asm round trip testing in ci the following job needs to be enabled for arm armlb ilrt this is a similar issue to dotnet runtime
1
7,341
6,910,955,776
IssuesEvent
2017-11-28 05:42:53
Khan/KaTeX
https://api.github.com/repos/Khan/KaTeX
closed
cache node_modules on travis-ci
infrastructure
Currently `npm install` takes about 25 seconds. Caching `node_modules` should be able to save some of that time. See https://blog.travis-ci.com/2016-05-03-caches-are-coming-to-everyone for details.
1.0
cache node_modules on travis-ci - Currently `npm install` takes about 25 seconds. Caching `node_modules` should be able to save some of that time. See https://blog.travis-ci.com/2016-05-03-caches-are-coming-to-everyone for details.
infrastructure
cache node modules on travis ci currently npm install takes about seconds caching node modules should be able to save some of that time see for details
1
675,125
23,079,826,085
IssuesEvent
2022-07-26 05:52:50
cub42d/cub3d
https://api.github.com/repos/cub42d/cub3d
opened
[To do] Bonus
Status: To Do Priority: High
## Description 간단한 보너스 구현 ## Progress - [ ] 스프라이트 물체 - [ ] 문 구현 (블록 삭제 생성 기반) - [ ] 미니맵 (방향 표시 안하고 미니맵만) - [ ] 마우스 기능 ## ETC 할수 있는 만큼만.....!!
1.0
[To do] Bonus - ## Description 간단한 보너스 구현 ## Progress - [ ] 스프라이트 물체 - [ ] 문 구현 (블록 삭제 생성 기반) - [ ] 미니맵 (방향 표시 안하고 미니맵만) - [ ] 마우스 기능 ## ETC 할수 있는 만큼만.....!!
non_infrastructure
bonus description 간단한 보너스 구현 progress 스프라이트 물체 문 구현 블록 삭제 생성 기반 미니맵 방향 표시 안하고 미니맵만 마우스 기능 etc 할수 있는 만큼만
0
34,959
2,789,572,780
IssuesEvent
2015-05-08 20:10:04
blacklocus/anvil
https://api.github.com/repos/blacklocus/anvil
opened
anvil show/hide series on a board
priority-high
by clicking the little squares in the legend here's a plugin https://github.com/markrcote/flot-hiddengraphs but it probably doesn't *solo* a series on double-click which would also be useful.
1.0
anvil show/hide series on a board - by clicking the little squares in the legend here's a plugin https://github.com/markrcote/flot-hiddengraphs but it probably doesn't *solo* a series on double-click which would also be useful.
non_infrastructure
anvil show hide series on a board by clicking the little squares in the legend here s a plugin but it probably doesn t solo a series on double click which would also be useful
0
10,578
8,637,982,215
IssuesEvent
2018-11-23 13:15:15
srinikoganti/sams
https://api.github.com/repos/srinikoganti/sams
closed
Infrastructure: The duplicate records in building, floor, room are not deleted.
Infrastructure P1
Infrastructure: The duplicate records in building, floor, room are not deleted. ![image](https://user-images.githubusercontent.com/33898419/48935435-b7bd3f80-ef2d-11e8-9cae-68147cd17351.png) ![image](https://user-images.githubusercontent.com/33898419/48935486-e4715700-ef2d-11e8-9030-47437c2eff77.png)
1.0
Infrastructure: The duplicate records in building, floor, room are not deleted. - Infrastructure: The duplicate records in building, floor, room are not deleted. ![image](https://user-images.githubusercontent.com/33898419/48935435-b7bd3f80-ef2d-11e8-9cae-68147cd17351.png) ![image](https://user-images.githubusercontent.com/33898419/48935486-e4715700-ef2d-11e8-9030-47437c2eff77.png)
infrastructure
infrastructure the duplicate records in building floor room are not deleted infrastructure the duplicate records in building floor room are not deleted
1
31,683
26,003,232,110
IssuesEvent
2022-12-20 16:56:30
10up/plugin-check
https://api.github.com/repos/10up/plugin-check
closed
Test Infrastructure with PHPUnit
[Type] Enhancement Infrastructure 7 Milestone 1
## Description Setup test infrastructure with PHPUnit ## Acceptance Criteria - Install the following packages with composer: - `"phpunit/phpunit"` - `"squizlabs/php_codesniffer"` - `"wp-coding-standards/wpcs"` - Create a `phpunit.xml.dist` file to run tests from the `/tests` folder - Create a `phpcs.xml.dist` file to run PHPCS tests - Update `composer.json` with the following `scripts`: - `"lint": "phpcs --standard=phpcs.xml.dist"` - to run PHPCS - `"test": "phpunit -c phpunit.xml.dist --verbose"` - to run PHPUnit tests - Install the following packages as dev dependencies with npm - `@wordpress/env` - `@wordpress/scripts` - Update `package.json` with the following `scripts`: - `"wp-env": "wp-env"` - `"lint-php": "wp-env run composer run-script lint"` - `"test-php": "wp-env run phpunit 'phpunit -c /var/www/html/wp-content/plugins/performance/phpunit.xml.dist --verbose'"` - `"lint-js": "wp-scripts lint-js ./bin"` - Create the following GitHub workflows in the `.github/workflows/` folder. (See [Performance Lab Plugin](https://github.com/WordPress/performance/tree/trunk/.github/workflows) for examples) - `js-lint.yml` - `php-lint.yml` - `php-test.yml`
1.0
Test Infrastructure with PHPUnit - ## Description Setup test infrastructure with PHPUnit ## Acceptance Criteria - Install the following packages with composer: - `"phpunit/phpunit"` - `"squizlabs/php_codesniffer"` - `"wp-coding-standards/wpcs"` - Create a `phpunit.xml.dist` file to run tests from the `/tests` folder - Create a `phpcs.xml.dist` file to run PHPCS tests - Update `composer.json` with the following `scripts`: - `"lint": "phpcs --standard=phpcs.xml.dist"` - to run PHPCS - `"test": "phpunit -c phpunit.xml.dist --verbose"` - to run PHPUnit tests - Install the following packages as dev dependencies with npm - `@wordpress/env` - `@wordpress/scripts` - Update `package.json` with the following `scripts`: - `"wp-env": "wp-env"` - `"lint-php": "wp-env run composer run-script lint"` - `"test-php": "wp-env run phpunit 'phpunit -c /var/www/html/wp-content/plugins/performance/phpunit.xml.dist --verbose'"` - `"lint-js": "wp-scripts lint-js ./bin"` - Create the following GitHub workflows in the `.github/workflows/` folder. (See [Performance Lab Plugin](https://github.com/WordPress/performance/tree/trunk/.github/workflows) for examples) - `js-lint.yml` - `php-lint.yml` - `php-test.yml`
infrastructure
test infrastructure with phpunit description setup test infrastructure with phpunit acceptance criteria install the following packages with composer phpunit phpunit squizlabs php codesniffer wp coding standards wpcs create a phpunit xml dist file to run tests from the tests folder create a phpcs xml dist file to run phpcs tests update composer json with the following scripts lint phpcs standard phpcs xml dist to run phpcs test phpunit c phpunit xml dist verbose to run phpunit tests install the following packages as dev dependencies with npm wordpress env wordpress scripts update package json with the following scripts wp env wp env lint php wp env run composer run script lint test php wp env run phpunit phpunit c var www html wp content plugins performance phpunit xml dist verbose lint js wp scripts lint js bin create the following github workflows in the github workflows folder see for examples js lint yml php lint yml php test yml
1
26,170
19,705,089,692
IssuesEvent
2022-01-12 20:57:37
muadev/tendercats
https://api.github.com/repos/muadev/tendercats
opened
Warnings silencionas en la action de fix-code
infrastructure
Tenemos un problema con nuestra acción de lint, porque estas warnings deberían aparecer como algo a corregir en el pull request: ![image](https://user-images.githubusercontent.com/142792/149220213-5cc8cc5d-f689-42da-9588-dd16772181a2.png) Tal vez se pueda hacer que salgan como errores, o algo así.
1.0
Warnings silencionas en la action de fix-code - Tenemos un problema con nuestra acción de lint, porque estas warnings deberían aparecer como algo a corregir en el pull request: ![image](https://user-images.githubusercontent.com/142792/149220213-5cc8cc5d-f689-42da-9588-dd16772181a2.png) Tal vez se pueda hacer que salgan como errores, o algo así.
infrastructure
warnings silencionas en la action de fix code tenemos un problema con nuestra acción de lint porque estas warnings deberían aparecer como algo a corregir en el pull request tal vez se pueda hacer que salgan como errores o algo así
1
28,502
23,298,280,994
IssuesEvent
2022-08-06 23:33:11
jrsmith3/ibei
https://api.github.com/repos/jrsmith3/ibei
closed
Write `tox` action to build documentation
development infrastructure
# Overview The scope of this ticket is to write a `tox` action that builds the package documentation. # Related issues * Depends on #48.
1.0
Write `tox` action to build documentation - # Overview The scope of this ticket is to write a `tox` action that builds the package documentation. # Related issues * Depends on #48.
infrastructure
write tox action to build documentation overview the scope of this ticket is to write a tox action that builds the package documentation related issues depends on
1
120,914
4,796,472,690
IssuesEvent
2016-11-01 08:01:00
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
[k8s.io] PreStop should call prestop when killing a pod [Conformance] {Kubernetes e2e suite}
kind/flake priority/P2
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gce-etcd3/1824/ Failed: [k8s.io] PreStop should call prestop when killing a pod [Conformance] {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pre_stop.go:167 validating pre-stop. Expected error: <*errors.errorString | 0xc420415820>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pre_stop.go:159 ``` Previous issues for this test: #30287
1.0
[k8s.io] PreStop should call prestop when killing a pod [Conformance] {Kubernetes e2e suite} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gce-etcd3/1824/ Failed: [k8s.io] PreStop should call prestop when killing a pod [Conformance] {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pre_stop.go:167 validating pre-stop. Expected error: <*errors.errorString | 0xc420415820>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/pre_stop.go:159 ``` Previous issues for this test: #30287
non_infrastructure
prestop should call prestop when killing a pod kubernetes suite failed prestop should call prestop when killing a pod kubernetes suite go src io kubernetes output dockerized go src io kubernetes test pre stop go validating pre stop expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test pre stop go previous issues for this test
0
156,777
5,989,089,485
IssuesEvent
2017-06-02 07:32:51
universAAL/ui
https://api.github.com/repos/universAAL/ui
closed
Send extra information within Submit button
help wanted imported priority 3
_Originally Opened: Hector Galan (2012-05-16 10:02:53_) _Originally Closed: 2013-09-18 19:38:30_ Let's suppose I am showing the nutritional menu for today. The menu has several dishes. I am able to show every dish using a SimpleOutput object. Next to it I place a submit button, one for each dish. When that button is pressed, it should load the recipe of that dish. Each button is built like this: Submit(Group parent, Label label, String id); Afterwards in the handleUIResponse method, I detect the id of the button pressed. Now, this has several problems. If there are 5 dishes, there will be 5 &quot;Show recipe&quot; buttons, each one with a custom ID like this: &quot;urn:samples.nutrition.uiclient:UIMenus#seeRecipeIDButton&quot;. Where IDButton is an integer that identifies the button. So I am using the submission ID of each button to send extra information such as the ID of the recipe to be loaded. My question is: is it possible to send or link the recipe ID or any other information to a button so that the handleUIResponse method is able to react accordingly? -- From: _this issue has been automatically imported from our old issue tracker_
1.0
Send extra information within Submit button - _Originally Opened: Hector Galan (2012-05-16 10:02:53_) _Originally Closed: 2013-09-18 19:38:30_ Let's suppose I am showing the nutritional menu for today. The menu has several dishes. I am able to show every dish using a SimpleOutput object. Next to it I place a submit button, one for each dish. When that button is pressed, it should load the recipe of that dish. Each button is built like this: Submit(Group parent, Label label, String id); Afterwards in the handleUIResponse method, I detect the id of the button pressed. Now, this has several problems. If there are 5 dishes, there will be 5 &quot;Show recipe&quot; buttons, each one with a custom ID like this: &quot;urn:samples.nutrition.uiclient:UIMenus#seeRecipeIDButton&quot;. Where IDButton is an integer that identifies the button. So I am using the submission ID of each button to send extra information such as the ID of the recipe to be loaded. My question is: is it possible to send or link the recipe ID or any other information to a button so that the handleUIResponse method is able to react accordingly? -- From: _this issue has been automatically imported from our old issue tracker_
non_infrastructure
send extra information within submit button originally opened hector galan originally closed let s suppose i am showing the nutritional menu for today the menu has several dishes i am able to show every dish using a simpleoutput object next to it i place a submit button one for each dish when that button is pressed it should load the recipe of that dish each button is built like this submit group parent label label string id afterwards in the handleuiresponse method i detect the id of the button pressed now this has several problems if there are dishes there will be quot show recipe quot buttons each one with a custom id like this quot urn samples nutrition uiclient uimenus seerecipeidbutton quot where idbutton is an integer that identifies the button so i am using the submission id of each button to send extra information such as the id of the recipe to be loaded my question is is it possible to send or link the recipe id or any other information to a button so that the handleuiresponse method is able to react accordingly from this issue has been automatically imported from our old issue tracker
0
19,700
13,397,032,749
IssuesEvent
2020-09-03 10:54:15
blockframes/blockframes
https://api.github.com/repos/blockframes/blockframes
closed
Reproduce prod environment on staging
Infrastructure
## Description Staging has diverged from prod and is no longer a trusted test environment. In order to test properly we need to have on staging an up-to-date firestore & firestorage taken from prod. We also need to anonymize prod firestore beforehand by changing: - Org name (`orgs.denomination.full` & `orgs.denomination.public`) - Org contact email address (`orgs.email`) - User name (`user.firstName` & `user.lastName`) - User email (`user.email`) Anonymizing prod protects users' data privacy and avoid staging to spam emails to real users. In order to be easily readable for us (and thus better testing) we used these conventions when generating the fake data: - org contact email address was derived from org name (contact@orgName.fake.cascade8.com) - user email was derived from user name (userName@orgName.fake.cascade8.com) These conventions should be updated to use real email address owned by cascade8 to avoid our Sendgrid email score to be polluted by the tests. E.g. use something like fakeAccounts+orgName-contact@cascade8.com & fakeAccounts+userName-orgName@cascade8.com The process should be this: 1. export prod db & storage 2. anonymize prod db 3. import anonymized prod db & storage 4. erase and recreate firebase authentication users based on anonymize prod db Best would be to be able to activate that script whenever we want and/or to activate it on each staging deploy. ## Current tools - Functions exist to backup & restore firestore following this url: projectHostingUrl/admin/data/backup & projectHostingUrl/admin/data/restore. [Here](https://www.notion.so/cascade8/Data-import-backup-c3279edb834f4a25b130adaecacef0d4) is an article to set it up on a project. When restore function is used, authentication is reset to match firestore - [Notion article](https://www.notion.so/cascade8/Bucket-storage-backup-restore-000b1dbb9df8453d807377e158a2450b) to use gs util to import & export firestorage; Google cloud also [exposes functions](https://console.cloud.google.com/transfer/cloud?project=blockframes&authuser=1) to do the same - Anonymizing prod was done with a Python script ## Acceptance criteria - [ ] Update backup & restore db functions if needed ([ressource](https://dev.to/mikgross/firebase-firestore-migrations-olg) from François - [ ] Add a storage backup & restore to the db functions - [ ] Include a script to anonymize db - [ ] Add it on the deploy staging flow and/or as a callable script
1.0
Reproduce prod environment on staging - ## Description Staging has diverged from prod and is no longer a trusted test environment. In order to test properly we need to have on staging an up-to-date firestore & firestorage taken from prod. We also need to anonymize prod firestore beforehand by changing: - Org name (`orgs.denomination.full` & `orgs.denomination.public`) - Org contact email address (`orgs.email`) - User name (`user.firstName` & `user.lastName`) - User email (`user.email`) Anonymizing prod protects users' data privacy and avoid staging to spam emails to real users. In order to be easily readable for us (and thus better testing) we used these conventions when generating the fake data: - org contact email address was derived from org name (contact@orgName.fake.cascade8.com) - user email was derived from user name (userName@orgName.fake.cascade8.com) These conventions should be updated to use real email address owned by cascade8 to avoid our Sendgrid email score to be polluted by the tests. E.g. use something like fakeAccounts+orgName-contact@cascade8.com & fakeAccounts+userName-orgName@cascade8.com The process should be this: 1. export prod db & storage 2. anonymize prod db 3. import anonymized prod db & storage 4. erase and recreate firebase authentication users based on anonymize prod db Best would be to be able to activate that script whenever we want and/or to activate it on each staging deploy. ## Current tools - Functions exist to backup & restore firestore following this url: projectHostingUrl/admin/data/backup & projectHostingUrl/admin/data/restore. [Here](https://www.notion.so/cascade8/Data-import-backup-c3279edb834f4a25b130adaecacef0d4) is an article to set it up on a project. When restore function is used, authentication is reset to match firestore - [Notion article](https://www.notion.so/cascade8/Bucket-storage-backup-restore-000b1dbb9df8453d807377e158a2450b) to use gs util to import & export firestorage; Google cloud also [exposes functions](https://console.cloud.google.com/transfer/cloud?project=blockframes&authuser=1) to do the same - Anonymizing prod was done with a Python script ## Acceptance criteria - [ ] Update backup & restore db functions if needed ([ressource](https://dev.to/mikgross/firebase-firestore-migrations-olg) from François - [ ] Add a storage backup & restore to the db functions - [ ] Include a script to anonymize db - [ ] Add it on the deploy staging flow and/or as a callable script
infrastructure
reproduce prod environment on staging description staging has diverged from prod and is no longer a trusted test environment in order to test properly we need to have on staging an up to date firestore firestorage taken from prod we also need to anonymize prod firestore beforehand by changing org name orgs denomination full orgs denomination public org contact email address orgs email user name user firstname user lastname user email user email anonymizing prod protects users data privacy and avoid staging to spam emails to real users in order to be easily readable for us and thus better testing we used these conventions when generating the fake data org contact email address was derived from org name contact orgname fake com user email was derived from user name username orgname fake com these conventions should be updated to use real email address owned by to avoid our sendgrid email score to be polluted by the tests e g use something like fakeaccounts orgname contact com fakeaccounts username orgname com the process should be this export prod db storage anonymize prod db import anonymized prod db storage erase and recreate firebase authentication users based on anonymize prod db best would be to be able to activate that script whenever we want and or to activate it on each staging deploy current tools functions exist to backup restore firestore following this url projecthostingurl admin data backup projecthostingurl admin data restore is an article to set it up on a project when restore function is used authentication is reset to match firestore to use gs util to import export firestorage google cloud also to do the same anonymizing prod was done with a python script acceptance criteria update backup restore db functions if needed from françois add a storage backup restore to the db functions include a script to anonymize db add it on the deploy staging flow and or as a callable script
1
22,577
15,285,735,995
IssuesEvent
2021-02-23 13:55:24
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Windows libraries test legs are timing out
area-Infrastructure-libraries untriaged
See for example here: ``` Waiting for completion of job cfbcdf1a-1496-4ceb-a6de-365182e0d3df on Windows.81.Amd64.Open Waiting for completion of job 75bcdab5-42a4-4787-ae3a-34db23d08327 on (Windows.Nano.1809.Amd64.Open)windows.10.amd64.serverrs5.open@mcr.microsoft.com/dotnet-buildtools/prereqs:nanoserver-1809-helix-amd64-08e8e40-20200107182504 Waiting for completion of job 0382b9cd-7899-4c1a-9d95-817f492f70ac on (Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Waiting for completion of job 809ae6cd-0127-44d0-911e-bc63f8f7342e on Windows.10.Amd64.Server19H1.Open Waiting for completion of job 7d564f0a-2775-4fd6-ae04-e69aed65699d on Windows.10.Amd64.ServerRS5.Open Job 7d564f0a-2775-4fd6-ae04-e69aed65699d on Windows.10.Amd64.ServerRS5.Open is completed with 264 finished work items. Job cfbcdf1a-1496-4ceb-a6de-365182e0d3df on Windows.81.Amd64.Open is completed with 264 finished work items. Job 0382b9cd-7899-4c1a-9d95-817f492f70ac on (Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c is completed with 264 finished work items. Job 75bcdab5-42a4-4787-ae3a-34db23d08327 on (Windows.Nano.1809.Amd64.Open)windows.10.amd64.serverrs5.open@mcr.microsoft.com/dotnet-buildtools/prereqs:nanoserver-1809-helix-amd64-08e8e40-20200107182504 is completed with 264 finished work items. ``` It's not always 19H1, it varies... cc @safern
1.0
Windows libraries test legs are timing out - See for example here: ``` Waiting for completion of job cfbcdf1a-1496-4ceb-a6de-365182e0d3df on Windows.81.Amd64.Open Waiting for completion of job 75bcdab5-42a4-4787-ae3a-34db23d08327 on (Windows.Nano.1809.Amd64.Open)windows.10.amd64.serverrs5.open@mcr.microsoft.com/dotnet-buildtools/prereqs:nanoserver-1809-helix-amd64-08e8e40-20200107182504 Waiting for completion of job 0382b9cd-7899-4c1a-9d95-817f492f70ac on (Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Waiting for completion of job 809ae6cd-0127-44d0-911e-bc63f8f7342e on Windows.10.Amd64.Server19H1.Open Waiting for completion of job 7d564f0a-2775-4fd6-ae04-e69aed65699d on Windows.10.Amd64.ServerRS5.Open Job 7d564f0a-2775-4fd6-ae04-e69aed65699d on Windows.10.Amd64.ServerRS5.Open is completed with 264 finished work items. Job cfbcdf1a-1496-4ceb-a6de-365182e0d3df on Windows.81.Amd64.Open is completed with 264 finished work items. Job 0382b9cd-7899-4c1a-9d95-817f492f70ac on (Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c is completed with 264 finished work items. Job 75bcdab5-42a4-4787-ae3a-34db23d08327 on (Windows.Nano.1809.Amd64.Open)windows.10.amd64.serverrs5.open@mcr.microsoft.com/dotnet-buildtools/prereqs:nanoserver-1809-helix-amd64-08e8e40-20200107182504 is completed with 264 finished work items. ``` It's not always 19H1, it varies... cc @safern
infrastructure
windows libraries test legs are timing out see for example here waiting for completion of job on windows open waiting for completion of job on windows nano open windows open mcr microsoft com dotnet buildtools prereqs nanoserver helix waiting for completion of job on windows server core open windows open mcr microsoft com dotnet buildtools prereqs windowsservercore helix waiting for completion of job on windows open waiting for completion of job on windows open job on windows open is completed with finished work items job on windows open is completed with finished work items job on windows server core open windows open mcr microsoft com dotnet buildtools prereqs windowsservercore helix is completed with finished work items job on windows nano open windows open mcr microsoft com dotnet buildtools prereqs nanoserver helix is completed with finished work items it s not always it varies cc safern
1
290,300
25,046,934,995
IssuesEvent
2022-11-05 11:27:55
nupac/nupac
https://api.github.com/repos/nupac/nupac
opened
Develop nushell testing framework and rewrite tests
refactor priority: CRITICAL github tests
### Describe the issue . ### Argumentation . ### Screenshots and other helpful media _No response_ ### Additional context _No response_
1.0
Develop nushell testing framework and rewrite tests - ### Describe the issue . ### Argumentation . ### Screenshots and other helpful media _No response_ ### Additional context _No response_
non_infrastructure
develop nushell testing framework and rewrite tests describe the issue argumentation screenshots and other helpful media no response additional context no response
0
77,216
26,854,229,259
IssuesEvent
2023-02-03 13:27:33
matrix-org/synapse
https://api.github.com/repos/matrix-org/synapse
closed
Logging in Rust code not working (in unit tests?)
S-Minor T-Defect O-Occasional
We use [pyo3-log](https://docs.rs/pyo3-log/latest/pyo3_log/) to get Rust logs into the Python logging framework; this seems to not work properly. If you put e.g. a `warn!(...)` then it doesn't print out anywhere. We initialize the logging module in `lib.rs`: https://github.com/matrix-org/synapse/blob/42d261c32f13e2de7494a0ade77c1f7b646af1fe/rust/src/lib.rs#L23 And we import the synapse rust code *very early* in the Synapse process: https://github.com/matrix-org/synapse/blob/dc37b68a25754240243cdca6f521919abfe71db0/synapse/__init__.py#L83 ---- I think we're importing the rust code before the loggers are configured, which then causes `pyo3-log` to initialize and [cache the Python logging configuration](https://docs.rs/pyo3-log/latest/pyo3_log/#performance-filtering-and-caching), then by the time we actually call Rust code we've reconfigured the python loggers, but that doesn't propagate into Rust. I ran into this in unittests, but I think it would happen during normal usage as well.
1.0
Logging in Rust code not working (in unit tests?) - We use [pyo3-log](https://docs.rs/pyo3-log/latest/pyo3_log/) to get Rust logs into the Python logging framework; this seems to not work properly. If you put e.g. a `warn!(...)` then it doesn't print out anywhere. We initialize the logging module in `lib.rs`: https://github.com/matrix-org/synapse/blob/42d261c32f13e2de7494a0ade77c1f7b646af1fe/rust/src/lib.rs#L23 And we import the synapse rust code *very early* in the Synapse process: https://github.com/matrix-org/synapse/blob/dc37b68a25754240243cdca6f521919abfe71db0/synapse/__init__.py#L83 ---- I think we're importing the rust code before the loggers are configured, which then causes `pyo3-log` to initialize and [cache the Python logging configuration](https://docs.rs/pyo3-log/latest/pyo3_log/#performance-filtering-and-caching), then by the time we actually call Rust code we've reconfigured the python loggers, but that doesn't propagate into Rust. I ran into this in unittests, but I think it would happen during normal usage as well.
non_infrastructure
logging in rust code not working in unit tests we use to get rust logs into the python logging framework this seems to not work properly if you put e g a warn then it doesn t print out anywhere we initialize the logging module in lib rs and we import the synapse rust code very early in the synapse process i think we re importing the rust code before the loggers are configured which then causes log to initialize and then by the time we actually call rust code we ve reconfigured the python loggers but that doesn t propagate into rust i ran into this in unittests but i think it would happen during normal usage as well
0
20,239
13,770,989,570
IssuesEvent
2020-10-07 21:09:46
grpc/grpc.io
https://api.github.com/repos/grpc/grpc.io
closed
Upgrade Hugo to v0.71.0+
e0-minutes e1-hours infrastructure p2-medium
We're currently using v0.70.0. Notable features in more recent version include: - [v0.71.0](https://gohugo.io/news/0.71.0-relnotes/): [Markdown Render Hooks](https://gohugo.io/getting-started/configuration-markup#markdown-render-hooks)
1.0
Upgrade Hugo to v0.71.0+ - We're currently using v0.70.0. Notable features in more recent version include: - [v0.71.0](https://gohugo.io/news/0.71.0-relnotes/): [Markdown Render Hooks](https://gohugo.io/getting-started/configuration-markup#markdown-render-hooks)
infrastructure
upgrade hugo to we re currently using notable features in more recent version include
1
10,698
8,687,480,705
IssuesEvent
2018-12-03 13:52:42
elastic/beats
https://api.github.com/repos/elastic/beats
opened
Add checks for the location of distributed dashboard files
:Packaging :infrastructure module
In #9307 it was reported that a dashboard was not being installed on setup, this was caused by a dashboard file placed in the wrong directory (#9313). Add a check to detect dashboard files distributed in wrong paths.
1.0
Add checks for the location of distributed dashboard files - In #9307 it was reported that a dashboard was not being installed on setup, this was caused by a dashboard file placed in the wrong directory (#9313). Add a check to detect dashboard files distributed in wrong paths.
infrastructure
add checks for the location of distributed dashboard files in it was reported that a dashboard was not being installed on setup this was caused by a dashboard file placed in the wrong directory add a check to detect dashboard files distributed in wrong paths
1
241,888
26,256,971,959
IssuesEvent
2023-01-06 02:12:27
faizulho/sanity-jigsaw-blog
https://api.github.com/repos/faizulho/sanity-jigsaw-blog
opened
CVE-2021-23436 (High) detected in immer-5.3.6.tgz
security vulnerability
## CVE-2021-23436 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-5.3.6.tgz</b></p></summary> <p>Create your next immutable state by mutating the current one</p> <p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-5.3.6.tgz">https://registry.npmjs.org/immer/-/immer-5.3.6.tgz</a></p> <p>Path to dependency file: /studio/package.json</p> <p>Path to vulnerable library: /studio/node_modules/immer/package.json</p> <p> Dependency Hierarchy: - desk-tool-1.150.8.tgz (Root Library) - form-builder-1.150.8.tgz - portable-text-editor-0.1.8.tgz - slate-0.58.4.tgz - :x: **immer-5.3.6.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type. <p>Publish Date: 2021-09-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23436>CVE-2021-23436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p> <p>Release Date: 2021-09-01</p> <p>Fix Resolution (immer): 9.0.6</p> <p>Direct dependency fix Resolution (@sanity/desk-tool): 2.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23436 (High) detected in immer-5.3.6.tgz - ## CVE-2021-23436 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-5.3.6.tgz</b></p></summary> <p>Create your next immutable state by mutating the current one</p> <p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-5.3.6.tgz">https://registry.npmjs.org/immer/-/immer-5.3.6.tgz</a></p> <p>Path to dependency file: /studio/package.json</p> <p>Path to vulnerable library: /studio/node_modules/immer/package.json</p> <p> Dependency Hierarchy: - desk-tool-1.150.8.tgz (Root Library) - form-builder-1.150.8.tgz - portable-text-editor-0.1.8.tgz - slate-0.58.4.tgz - :x: **immer-5.3.6.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type. <p>Publish Date: 2021-09-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23436>CVE-2021-23436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p> <p>Release Date: 2021-09-01</p> <p>Fix Resolution (immer): 9.0.6</p> <p>Direct dependency fix Resolution (@sanity/desk-tool): 2.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in immer tgz cve high severity vulnerability vulnerable library immer tgz create your next immutable state by mutating the current one library home page a href path to dependency file studio package json path to vulnerable library studio node modules immer package json dependency hierarchy desk tool tgz root library form builder tgz portable text editor tgz slate tgz x immer tgz vulnerable library found in base branch main vulnerability details this affects the package immer before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays in particular this bypass is possible because the condition p proto p constructor in applypatches returns false if p is or the operator strict equality operator returns false if the operands have different type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution immer direct dependency fix resolution sanity desk tool step up your open source security game with mend
0
174,439
21,300,067,505
IssuesEvent
2022-04-15 01:03:08
shaimael/easybuggy2
https://api.github.com/repos/shaimael/easybuggy2
closed
CVE-2018-1313 (Medium) detected in derby-10.8.3.0.jar - autoclosed
security vulnerability
## CVE-2018-1313 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>derby-10.8.3.0.jar</b></p></summary> <p>Contains the core Apache Derby database engine, which also includes the embedded JDBC driver.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /repository/org/apache/derby/derby/10.8.3.0/derby-10.8.3.0.jar</p> <p> Dependency Hierarchy: - :x: **derby-10.8.3.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/easybuggy2/commit/ce89fd7428395b7b7deb3984e6d7f139611ecc6f">ce89fd7428395b7b7deb3984e6d7f139611ecc6f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user's control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work. <p>Publish Date: 2018-05-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313>CVE-2018-1313</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313</a></p> <p>Release Date: 2018-05-07</p> <p>Fix Resolution: org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.derby","packageName":"derby","packageVersion":"10.8.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.derby:derby:10.8.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1313","vulnerabilityDetails":"In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user\u0027s control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1313 (Medium) detected in derby-10.8.3.0.jar - autoclosed - ## CVE-2018-1313 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>derby-10.8.3.0.jar</b></p></summary> <p>Contains the core Apache Derby database engine, which also includes the embedded JDBC driver.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /repository/org/apache/derby/derby/10.8.3.0/derby-10.8.3.0.jar</p> <p> Dependency Hierarchy: - :x: **derby-10.8.3.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/easybuggy2/commit/ce89fd7428395b7b7deb3984e6d7f139611ecc6f">ce89fd7428395b7b7deb3984e6d7f139611ecc6f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user's control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work. <p>Publish Date: 2018-05-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313>CVE-2018-1313</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313</a></p> <p>Release Date: 2018-05-07</p> <p>Fix Resolution: org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.derby","packageName":"derby","packageVersion":"10.8.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.derby:derby:10.8.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1313","vulnerabilityDetails":"In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user\u0027s control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve medium detected in derby jar autoclosed cve medium severity vulnerability vulnerable library derby jar contains the core apache derby database engine which also includes the embedded jdbc driver path to dependency file pom xml path to vulnerable library repository org apache derby derby derby jar dependency hierarchy x derby jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache derby to a specially crafted network packet can be used to request the derby network server to boot a database whose location and contents are under the user s control if the derby network server is not running with a java security manager policy file the attack is successful if the server is using a policy file the policy file must permit the database location to be read for the attack to work the default derby network server policy file distributed with the affected releases includes a permissive policy as the default network server policy which allows the attack to work publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache derby derbynet org apache derby derby org apache derby derbytools org apache derby derbyclient isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache derby derby isminimumfixversionavailable true minimumfixversion org apache derby derbynet org apache derby derby org apache derby derbytools org apache derby derbyclient isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache derby to a specially crafted network packet can be used to request the derby network server to boot a database whose location and contents are under the user control if the derby network server is not running with a java security manager policy file the attack is successful if the server is using a policy file the policy file must permit the database location to be read for the attack to work the default derby network server policy file distributed with the affected releases includes a permissive policy as the default network server policy which allows the attack to work vulnerabilityurl
0
14,898
2,831,390,190
IssuesEvent
2015-05-24 15:55:02
nobodyguy/dslrdashboard
https://api.github.com/repos/nobodyguy/dslrdashboard
closed
no live d3200
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1used usb wired link nojoy. 2. Used nikon wifi link no joy 3. What is the expected output? What do you see instead? Just see black screen What version of the product are you using? On what operating system? 30.22 on galaxy tab 4.0 Please provide any additional information below. Can take pictures and can focus but no liveview. ``` Original issue reported on code.google.com by `andy.eve...@gmail.com` on 21 Jul 2013 at 2:46
1.0
no live d3200 - ``` What steps will reproduce the problem? 1used usb wired link nojoy. 2. Used nikon wifi link no joy 3. What is the expected output? What do you see instead? Just see black screen What version of the product are you using? On what operating system? 30.22 on galaxy tab 4.0 Please provide any additional information below. Can take pictures and can focus but no liveview. ``` Original issue reported on code.google.com by `andy.eve...@gmail.com` on 21 Jul 2013 at 2:46
non_infrastructure
no live what steps will reproduce the problem usb wired link nojoy used nikon wifi link no joy what is the expected output what do you see instead just see black screen what version of the product are you using on what operating system on galaxy tab please provide any additional information below can take pictures and can focus but no liveview original issue reported on code google com by andy eve gmail com on jul at
0
33,991
28,070,689,754
IssuesEvent
2023-03-29 18:49:49
MinaProtocol/mina
https://api.github.com/repos/MinaProtocol/mina
closed
Current State of helm + terraform in develop
mainNet infrastructure
Largely these issues relate to integration tests, but its also just the integration tests doing their job and showing that part of the integration is broken. - [x] Nodes not connecting to the k8s seed, could be related to an issue introduced in the cluster over the weekend, could be related to CONNECT_PRIVATE_IPS env var - [x] runtimeConfig not working the way integration tests are using it - [x] -generate-genesis-proof=true should be possible to configure in helm independent of the runtimeConfig variable - [x] Static PeerIDs in a dynamic peers.txt list (@AryehAlexander is picking this up) - [x] Master should not build every hour, only every commit (or even every tag). If we do want to re-run master frequently it needs to only release new packages when manually triggered. - [x] Add `sudo -n` so that sudo doesn't hang when prompting for a password - [x] Remove need for `sudo chmod` in the one-click deploy steps because its not portable enough.
1.0
Current State of helm + terraform in develop - Largely these issues relate to integration tests, but its also just the integration tests doing their job and showing that part of the integration is broken. - [x] Nodes not connecting to the k8s seed, could be related to an issue introduced in the cluster over the weekend, could be related to CONNECT_PRIVATE_IPS env var - [x] runtimeConfig not working the way integration tests are using it - [x] -generate-genesis-proof=true should be possible to configure in helm independent of the runtimeConfig variable - [x] Static PeerIDs in a dynamic peers.txt list (@AryehAlexander is picking this up) - [x] Master should not build every hour, only every commit (or even every tag). If we do want to re-run master frequently it needs to only release new packages when manually triggered. - [x] Add `sudo -n` so that sudo doesn't hang when prompting for a password - [x] Remove need for `sudo chmod` in the one-click deploy steps because its not portable enough.
infrastructure
current state of helm terraform in develop largely these issues relate to integration tests but its also just the integration tests doing their job and showing that part of the integration is broken nodes not connecting to the seed could be related to an issue introduced in the cluster over the weekend could be related to connect private ips env var runtimeconfig not working the way integration tests are using it generate genesis proof true should be possible to configure in helm independent of the runtimeconfig variable static peerids in a dynamic peers txt list aryehalexander is picking this up master should not build every hour only every commit or even every tag if we do want to re run master frequently it needs to only release new packages when manually triggered add sudo n so that sudo doesn t hang when prompting for a password remove need for sudo chmod in the one click deploy steps because its not portable enough
1
443,240
30,880,889,755
IssuesEvent
2023-08-03 17:29:48
snivilised/arcadia
https://api.github.com/repos/snivilised/arcadia
closed
update branch protection setup info in readme
documentation
looks like the UI in github has changed, so the instructions need to be updated to match.
1.0
update branch protection setup info in readme - looks like the UI in github has changed, so the instructions need to be updated to match.
non_infrastructure
update branch protection setup info in readme looks like the ui in github has changed so the instructions need to be updated to match
0
6,883
6,654,486,233
IssuesEvent
2017-09-29 13:02:09
ManickYoj/eve-merchant
https://api.github.com/repos/ManickYoj/eve-merchant
opened
Migrate Dynamic Information to SQLite
Infrastructure Enhancemnt
To finish the move over to SQLite, we need to be able to read and write the API-gathered data to the database.
1.0
Migrate Dynamic Information to SQLite - To finish the move over to SQLite, we need to be able to read and write the API-gathered data to the database.
infrastructure
migrate dynamic information to sqlite to finish the move over to sqlite we need to be able to read and write the api gathered data to the database
1
23,303
16,046,677,319
IssuesEvent
2021-04-22 14:21:45
emory-libraries/blacklight-catalog
https://api.github.com/repos/emory-libraries/blacklight-catalog
closed
Configure Additional CloudWatch Alerting
Epic Infrastructure Story
**STORY** As part of an earlier spike, we have identified available disk space as a key metric to monitor and alert on. For further details see #348 SPIKE: Determine what CloudWatch Metrics are available to set benchmarks on SOLR Performance. **ACCEPTANCE** The following statistics are being monitored (benchmarked) with alerts set as follows: - [x] Solr disk utilization > 80% - #396 - [x] Web server disk utilization > 50% - #397 - [ ] JVM Heap Usage > 95% - 398 - [x] Solr system memory >= 100% - #399 - [x] Web server system memory >= 100% - #400 - [x] Solr CPU utilization > 90% - #401 - [x] Web server CPU utilization > 90% - #402 Notes: - [Prior work on DLP monitoring for various applications](https://app.zenhub.com/workspaces/digital-library-project-5bf484ae4b5806bc2bf6875b/issues/emory-libraries/dlp-curate/956)
1.0
Configure Additional CloudWatch Alerting - **STORY** As part of an earlier spike, we have identified available disk space as a key metric to monitor and alert on. For further details see #348 SPIKE: Determine what CloudWatch Metrics are available to set benchmarks on SOLR Performance. **ACCEPTANCE** The following statistics are being monitored (benchmarked) with alerts set as follows: - [x] Solr disk utilization > 80% - #396 - [x] Web server disk utilization > 50% - #397 - [ ] JVM Heap Usage > 95% - 398 - [x] Solr system memory >= 100% - #399 - [x] Web server system memory >= 100% - #400 - [x] Solr CPU utilization > 90% - #401 - [x] Web server CPU utilization > 90% - #402 Notes: - [Prior work on DLP monitoring for various applications](https://app.zenhub.com/workspaces/digital-library-project-5bf484ae4b5806bc2bf6875b/issues/emory-libraries/dlp-curate/956)
infrastructure
configure additional cloudwatch alerting story as part of an earlier spike we have identified available disk space as a key metric to monitor and alert on for further details see spike determine what cloudwatch metrics are available to set benchmarks on solr performance acceptance the following statistics are being monitored benchmarked with alerts set as follows solr disk utilization web server disk utilization jvm heap usage solr system memory web server system memory solr cpu utilization web server cpu utilization notes
1
34,687
30,279,338,660
IssuesEvent
2023-07-08 00:00:57
evcc-io/evcc
https://api.github.com/repos/evcc-io/evcc
closed
Reduce logging
enhancement infrastructure stale
``` [lp-1 ] DEBUG 2023/06/12 09:51:00 charge power: 2661W [site ] DEBUG 2023/06/12 09:51:00 planner tariff [site ] DEBUG 2023/06/12 09:51:00 pv power: 3328W [site ] DEBUG 2023/06/12 09:51:01 grid power: 89W [site ] DEBUG 2023/06/12 09:51:01 site power: 89W [lp-1 ] DEBUG 2023/06/12 09:51:01 charge currents: [11.8 0.033 0.033]A [lp-1 ] DEBUG 2023/06/12 09:51:01 detected active phases: 1p [lp-1 ] DEBUG 2023/06/12 09:51:01 charge meter total import: 1847.560kWh [lp-1 ] DEBUG 2023/06/12 09:51:01 charger status: C [lp-1 ] DEBUG 2023/06/12 09:51:01 vehicle soc: 79% [lp-1 ] DEBUG 2023/06/12 09:51:01 vehicle range: 230km [lp-1 ] DEBUG 2023/06/12 09:51:01 pv charge current: 11.5A = 11.9A + -0.389A (89W @ 1p) ``` These lines don't fit and are repeated per cycle: ``` [lp-1 ] DEBUG 2023/06/12 09:51:01 detected active phases: 1p [lp-1 ] DEBUG 2023/06/12 09:51:01 charge meter total import: 1847.560kWh ```
1.0
Reduce logging - ``` [lp-1 ] DEBUG 2023/06/12 09:51:00 charge power: 2661W [site ] DEBUG 2023/06/12 09:51:00 planner tariff [site ] DEBUG 2023/06/12 09:51:00 pv power: 3328W [site ] DEBUG 2023/06/12 09:51:01 grid power: 89W [site ] DEBUG 2023/06/12 09:51:01 site power: 89W [lp-1 ] DEBUG 2023/06/12 09:51:01 charge currents: [11.8 0.033 0.033]A [lp-1 ] DEBUG 2023/06/12 09:51:01 detected active phases: 1p [lp-1 ] DEBUG 2023/06/12 09:51:01 charge meter total import: 1847.560kWh [lp-1 ] DEBUG 2023/06/12 09:51:01 charger status: C [lp-1 ] DEBUG 2023/06/12 09:51:01 vehicle soc: 79% [lp-1 ] DEBUG 2023/06/12 09:51:01 vehicle range: 230km [lp-1 ] DEBUG 2023/06/12 09:51:01 pv charge current: 11.5A = 11.9A + -0.389A (89W @ 1p) ``` These lines don't fit and are repeated per cycle: ``` [lp-1 ] DEBUG 2023/06/12 09:51:01 detected active phases: 1p [lp-1 ] DEBUG 2023/06/12 09:51:01 charge meter total import: 1847.560kWh ```
infrastructure
reduce logging debug charge power debug planner tariff debug pv power debug grid power debug site power debug charge currents a debug detected active phases debug charge meter total import debug charger status c debug vehicle soc debug vehicle range debug pv charge current these lines don t fit and are repeated per cycle debug detected active phases debug charge meter total import
1
5,470
5,681,651,668
IssuesEvent
2017-04-13 07:18:21
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
opened
Lock npm versions
Infrastructure
**Describe the issue you encountered:** Lock all npm packages to the specific version, so that we are all working on the same dependencies. We should update all dependencies with every release. cc @bbondy @alexwykoff @bsclifton
1.0
Lock npm versions - **Describe the issue you encountered:** Lock all npm packages to the specific version, so that we are all working on the same dependencies. We should update all dependencies with every release. cc @bbondy @alexwykoff @bsclifton
infrastructure
lock npm versions describe the issue you encountered lock all npm packages to the specific version so that we are all working on the same dependencies we should update all dependencies with every release cc bbondy alexwykoff bsclifton
1
2,237
3,588,245,536
IssuesEvent
2016-01-30 21:53:40
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
TinCan xAPI Statement Structure
infrastructure Tincan / XAPI
This is to provide a base for discussion around how xAPI statements should be structured in relation to the ELMSLN course network and detail any potential issues. I'll provide a starting point with an example statement from a H5P activity in our deployment: ![image](https://cloud.githubusercontent.com/assets/4973810/10819586/cd6105c6-7e40-11e5-8bfc-b65a00e1ff18.png) I'll step through where the different values come from for each area and some of the issues I potentially have with these. ### Actor ``` { "name": "Cheryl Cole-849" "mbox": "mailto:cheryl.cole@example.com" } ``` The name is the username of the Drupal user with their UID appended to the end, and the "mbox" is the email address of the user. "mbox" is a type of Inverse Functional Identifier with the specification stating: - An Agent MUST be identified by one (1) of the four types of Inverse Functional Identifiers (see 4.1.2.3 Inverse Functional Identifier); - An Agent MUST NOT include more than one (1) Inverse Functional Identifier; There are [other options](https://github.com/adlnet/xAPI-Spec/blob/master/xAPI.md#inversefunctional) but It's difficult being restricted to one IFI (although I understand why they've done that). I'm not satisfied with how we've set it up at the moment, as our students can actually change their email address in the dashboard so "mbox" in this regard is a bad choice (and why we've also added Drupal UID to end of the name), but limiting the ID to one that's specific to Drupal or any other internal LMS seems to go against the principles of a student's data being transferable and being owned by them. ### Verb Verbs are kind of determined already by the H5P output, but as we spread this out more throughout the network we should make sure we use [TinCan recipes](https://tincanapi.com/recipes/) to make sure statements are consistent for certain activity types. ### Context ``` { "platform": "Customer Service Course for Delivery Drivers" } ``` This shows which course the statement took place within. This isn't actually where I wanted this information to sit but it's ended up that way. I actually wanted the following as I think this more accurately represents the spec: ``` "context": { { "grouping" : [ {"id" : "http://courses.buttercupstraining.co.uk/csdeliv"} ] } } ``` But there's other [context options here](https://github.com/adlnet/xAPI-Spec/blob/master/xAPI.md#context) where we could pass things like Section ID (section access string?), Service type etc. so these should be decided ### Object ``` "object": { "objectType": "Activity", "id": "http://courses.buttercupstraining.co.uk/csdeliv/node/44", "definition": { "name": { "en-UK": "Unit 6 Homework Questions" } } } ``` Here this picks out the URL that the H5P has been embedded on as the "id" and gives the "name" as the node title. One problem I have with this is that it almost prevents two learning activities being added to the same page, as both would have the same "id". There somehow needs to be a reference back to the original H5P instance of this in Media (or wherever we're embedding from). Hopefully this will kick off a good discussion for how xAPI will work within the ELMSLN environment. Happy to spin off any of these points into their own issue if required
1.0
TinCan xAPI Statement Structure - This is to provide a base for discussion around how xAPI statements should be structured in relation to the ELMSLN course network and detail any potential issues. I'll provide a starting point with an example statement from a H5P activity in our deployment: ![image](https://cloud.githubusercontent.com/assets/4973810/10819586/cd6105c6-7e40-11e5-8bfc-b65a00e1ff18.png) I'll step through where the different values come from for each area and some of the issues I potentially have with these. ### Actor ``` { "name": "Cheryl Cole-849" "mbox": "mailto:cheryl.cole@example.com" } ``` The name is the username of the Drupal user with their UID appended to the end, and the "mbox" is the email address of the user. "mbox" is a type of Inverse Functional Identifier with the specification stating: - An Agent MUST be identified by one (1) of the four types of Inverse Functional Identifiers (see 4.1.2.3 Inverse Functional Identifier); - An Agent MUST NOT include more than one (1) Inverse Functional Identifier; There are [other options](https://github.com/adlnet/xAPI-Spec/blob/master/xAPI.md#inversefunctional) but It's difficult being restricted to one IFI (although I understand why they've done that). I'm not satisfied with how we've set it up at the moment, as our students can actually change their email address in the dashboard so "mbox" in this regard is a bad choice (and why we've also added Drupal UID to end of the name), but limiting the ID to one that's specific to Drupal or any other internal LMS seems to go against the principles of a student's data being transferable and being owned by them. ### Verb Verbs are kind of determined already by the H5P output, but as we spread this out more throughout the network we should make sure we use [TinCan recipes](https://tincanapi.com/recipes/) to make sure statements are consistent for certain activity types. ### Context ``` { "platform": "Customer Service Course for Delivery Drivers" } ``` This shows which course the statement took place within. This isn't actually where I wanted this information to sit but it's ended up that way. I actually wanted the following as I think this more accurately represents the spec: ``` "context": { { "grouping" : [ {"id" : "http://courses.buttercupstraining.co.uk/csdeliv"} ] } } ``` But there's other [context options here](https://github.com/adlnet/xAPI-Spec/blob/master/xAPI.md#context) where we could pass things like Section ID (section access string?), Service type etc. so these should be decided ### Object ``` "object": { "objectType": "Activity", "id": "http://courses.buttercupstraining.co.uk/csdeliv/node/44", "definition": { "name": { "en-UK": "Unit 6 Homework Questions" } } } ``` Here this picks out the URL that the H5P has been embedded on as the "id" and gives the "name" as the node title. One problem I have with this is that it almost prevents two learning activities being added to the same page, as both would have the same "id". There somehow needs to be a reference back to the original H5P instance of this in Media (or wherever we're embedding from). Hopefully this will kick off a good discussion for how xAPI will work within the ELMSLN environment. Happy to spin off any of these points into their own issue if required
infrastructure
tincan xapi statement structure this is to provide a base for discussion around how xapi statements should be structured in relation to the elmsln course network and detail any potential issues i ll provide a starting point with an example statement from a activity in our deployment i ll step through where the different values come from for each area and some of the issues i potentially have with these actor name cheryl cole mbox mailto cheryl cole example com the name is the username of the drupal user with their uid appended to the end and the mbox is the email address of the user mbox is a type of inverse functional identifier with the specification stating an agent must be identified by one of the four types of inverse functional identifiers see inverse functional identifier an agent must not include more than one inverse functional identifier there are but it s difficult being restricted to one ifi although i understand why they ve done that i m not satisfied with how we ve set it up at the moment as our students can actually change their email address in the dashboard so mbox in this regard is a bad choice and why we ve also added drupal uid to end of the name but limiting the id to one that s specific to drupal or any other internal lms seems to go against the principles of a student s data being transferable and being owned by them verb verbs are kind of determined already by the output but as we spread this out more throughout the network we should make sure we use to make sure statements are consistent for certain activity types context platform customer service course for delivery drivers this shows which course the statement took place within this isn t actually where i wanted this information to sit but it s ended up that way i actually wanted the following as i think this more accurately represents the spec context grouping id but there s other where we could pass things like section id section access string service type etc so these should be decided object object objecttype activity id definition name en uk unit homework questions here this picks out the url that the has been embedded on as the id and gives the name as the node title one problem i have with this is that it almost prevents two learning activities being added to the same page as both would have the same id there somehow needs to be a reference back to the original instance of this in media or wherever we re embedding from hopefully this will kick off a good discussion for how xapi will work within the elmsln environment happy to spin off any of these points into their own issue if required
1
225,096
17,792,997,669
IssuesEvent
2021-08-31 18:28:41
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Manual test run on Android Tab for 1.29.x - Release
tests QA/Yes release-notes/exclude OS/Android ARM
## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Startup - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler` or `Wireshark` (or a similar application) ## Visual look - [x] Make sure thereafter every merge - [x] No Chrome/Chromium words appear on `brave://version` - [x] No Chrome/Chromium words appear on normal or private tabs - [x] No Chrome/Chromium words appear in site settings for `Location` / `Camera` / `Microphone` / `Augmented Reality` - [x] No Chrome/Chromium icons are shown in normal or private tabs ## Data Pre-Requisite: Put previous build shortcut on the home screen. Also, have several sites 'Added to home screen' (from 3 dots menu) and then upgrade to new build - [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, etc) - [x] Verify that the cookies from the previous build are preserved after upgrade - [x] Verify shortcut is still available on the home screen after upgrade - [x] Verify sites added to home screen are still visible and able to be used after upgrade - [x] Verify sync chain created in the previous version is still retained on upgrade - [x] Verify settings changes done in the previous version are still retained on upgrade - [x] Verify IPFS, ENS & Unstoppable Domain settings is retained after upgrade - [x] Verify tab-group setting doesn't reset upon upgrade - [x] Verify Clear Data on exit setting state from previous build is retained upon upgrade ## Custom tabs - [x] Make sure Brave handles links from Gmail, Slack - [x] Ensure custom tabs work even with sync enabled/disabled ## Developer Tools - [x] Verify you can inspect sub-links via dev tools ## Clear Data - [x] Verify Clear Data on exit works as intended ## Settings and Bottom bar - [x] Verify changing default settings are retained and don't cause the browser to crash - [ ] Verify bottom bar buttons (Home/Bookmark/Search/Tabs) work as expected ## IPFS, ENS & Unstoppable Domain - [x] Verify settings for IPFS, ENS & Unstoppable Domain show up under Brave Shields & Privacy settings ## Downloads - [x] Verify downloading a file works and that all actions on the download item work. - [x] Verify that PDF is downloaded over HTTPS at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is downloaded over HTTP at `http://www.pdf995.com/samples/pdf.pdf` ## Bravery settings - [x] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/` - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to `https://https-everywhere.badssl.com/` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that clicking through a cert error in `https://badssl.com/` works - [x] Visit `https://brianbondy.com/` and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. - [x] Verify that default Bravery settings take effect on pages with no site settings - [x] Verify that 3rd party storage results are blank at `https://jsfiddle.net/7ke9r14a/7/` when 3rd party cookies are blocked - [x] Choose a DNS provider from the list in Settings | Privacy | Use Secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear. - [x] Run all six test configurations on `https://dev-pages.brave.software/storage/ephemeral-storage.html` and confirm results are as expected per each test listed. ### Fingerprint Tests - [x] Visit `https://browserleaks.com/webrtc`, ensure 2 blocked items are listed in shields - [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy` ## Content Tests - [x] Go to `https://brianbondy.com/` and click on the twitter icon on the top right. Verify that context menus work in the new twitter tab - [x] Go to `https://trac.torproject.org/projects/tor/login` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again - [x] Open a GitHub issue and type some misspellings, make sure they aren't autocorrected - [x] Open an email on `http://mail.google.com/` or inbox.google.com and click on a link. Make sure it works - [x] Verify that `https://mixed-script.badssl.com/` shows up as grey not red (no mixed content scripts are run) ## Brave Rewards/Ads - [x] Verify wallet is auto-created after enabling rewards(either via Panel or Rewards page) - [x] Verify account balance shows correct BAT and USD value - [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [x] Verify grant details are shown in expanded view when a grant is claimed - [x] Verify monthly budget shows correct BAT and USD value - [x] Verify you can exclude a publisher from the auto-contribute table by clicking on the trash bin icon in the auto-contribute table - [x] Verify you can exclude a publisher by using the toggle on the Rewards Panel - [x] Verify you can remove excluded sites via `Restore All` button - [x] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %) - [x] Verify when you click on `Send a tip`, the custom tip banner displays - [x] Verify you can make a one-time tip and they display in tips panel - [x] Verify you can make a recurring tip and they display in tips panel - [x] Verify you can tip a verified publisher - [x] Verify you can tip a verified YouTube creator - [x] Verify tip panel shows a verified check mark for a verified publisher/verified YouTube creator - [x] Verify tip panel shows a message about the unverified publisher - [x] Verify BR panel shows the message about an unverified publisher - [x] Verify you can perform a contribution - [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [x] Verify that disabling Rewards and enabling it again does not lose state - [x] Verify that disabling auto-contribute and enabling it again does not lose state - [x] Verify unchecking `Allow contribution to videos` option doesn't list any YouTube creator in ac list - [x] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [x] Verify you can reset rewards from advance setting. Resetting should delete wallet and bring it back to the pre-optin state - [x] Verify on BR panel `Verify wallet` button loads verify wallet page when balance in >=15 BAT. - [x] Verify user still able to connect to existing Uphold account via panel even when balance is < 15 BAT - [x] Verify you are able to disconnect/re-connect a user wallet when min balance is available - [x] Upgrade from an older version - [x] Verify the wallet balance (if available) is retained - [x] Verify auto-contribute list is not lost after upgrade - [x] Verify tips list is not lost after upgrade - [x] Verify wallet panel transactions list is not lost after upgrade ### Brave Ads - [x] Verify ads is auto-enabled when rewards is enabled for the supported region - [x] Verify ads are only shown when the app is being used - [x] Verify ad notification are shown based on ads per hour setting - [x] Verify ad notifications stack up in notification tray - [x] Verify swipe left/right dismisses the ad notification when shown and is not stored in the notification tray - [x] Verify clicking on an ad notification shows the landing page - [x] Verify `view`,`clicked` and `landed` and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join sync chain by scanning the QR code - [x] Verify you are able to join sync chain using code words - [x] Verify you are able to create a sync chain on the device and add other devices to the chain via QR code/Code words - [x] Verify once sync chain is created, `Categories` option is shown in devices list - [x] Verify only `Bookmarks` is enabled in `Categories` by default - [x] Verify enabling `Sync everything` enables all other switches - [x] Verify existing bookmarks before joining sync chain also gets sync'd to all devices on the sync chain - [x] Verify `Autofill` data is sync'd to all devices when enabled - [x] Verify `History` is sync'd to all devices - [x] Verify `Open tab` form other devices shows up under history below the device name - [x] Verify `Password` is sync'd to all devices - [x] Verify `Settings` sync's site settings on all devices - [x] Verify sync works on an upgrade profile and new bookmarks added post-upgrade sync's across devices on the chain - [x] Verify adding a bookmark on custom tab gets sync'd across all devices in the chain - [x] Verify you are able to create a standalone sync chain with one device - [x] Verify `Remove this device` shows a confirmation alert before removing the sync chain on the device ## Top sites view - [x] Long-press on top sites to get to deletion mode, and delete a top site (note this will stop that site from showing up again on top sites, so you may not want to do this a site you want to keep there) ## Session storage - [x] Verify that tabs restore when closed, including active tab
1.0
Manual test run on Android Tab for 1.29.x - Release - ## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Startup - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler` or `Wireshark` (or a similar application) ## Visual look - [x] Make sure thereafter every merge - [x] No Chrome/Chromium words appear on `brave://version` - [x] No Chrome/Chromium words appear on normal or private tabs - [x] No Chrome/Chromium words appear in site settings for `Location` / `Camera` / `Microphone` / `Augmented Reality` - [x] No Chrome/Chromium icons are shown in normal or private tabs ## Data Pre-Requisite: Put previous build shortcut on the home screen. Also, have several sites 'Added to home screen' (from 3 dots menu) and then upgrade to new build - [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, etc) - [x] Verify that the cookies from the previous build are preserved after upgrade - [x] Verify shortcut is still available on the home screen after upgrade - [x] Verify sites added to home screen are still visible and able to be used after upgrade - [x] Verify sync chain created in the previous version is still retained on upgrade - [x] Verify settings changes done in the previous version are still retained on upgrade - [x] Verify IPFS, ENS & Unstoppable Domain settings is retained after upgrade - [x] Verify tab-group setting doesn't reset upon upgrade - [x] Verify Clear Data on exit setting state from previous build is retained upon upgrade ## Custom tabs - [x] Make sure Brave handles links from Gmail, Slack - [x] Ensure custom tabs work even with sync enabled/disabled ## Developer Tools - [x] Verify you can inspect sub-links via dev tools ## Clear Data - [x] Verify Clear Data on exit works as intended ## Settings and Bottom bar - [x] Verify changing default settings are retained and don't cause the browser to crash - [ ] Verify bottom bar buttons (Home/Bookmark/Search/Tabs) work as expected ## IPFS, ENS & Unstoppable Domain - [x] Verify settings for IPFS, ENS & Unstoppable Domain show up under Brave Shields & Privacy settings ## Downloads - [x] Verify downloading a file works and that all actions on the download item work. - [x] Verify that PDF is downloaded over HTTPS at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is downloaded over HTTP at `http://www.pdf995.com/samples/pdf.pdf` ## Bravery settings - [x] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/` - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to `https://https-everywhere.badssl.com/` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that clicking through a cert error in `https://badssl.com/` works - [x] Visit `https://brianbondy.com/` and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. - [x] Verify that default Bravery settings take effect on pages with no site settings - [x] Verify that 3rd party storage results are blank at `https://jsfiddle.net/7ke9r14a/7/` when 3rd party cookies are blocked - [x] Choose a DNS provider from the list in Settings | Privacy | Use Secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear. - [x] Run all six test configurations on `https://dev-pages.brave.software/storage/ephemeral-storage.html` and confirm results are as expected per each test listed. ### Fingerprint Tests - [x] Visit `https://browserleaks.com/webrtc`, ensure 2 blocked items are listed in shields - [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy` ## Content Tests - [x] Go to `https://brianbondy.com/` and click on the twitter icon on the top right. Verify that context menus work in the new twitter tab - [x] Go to `https://trac.torproject.org/projects/tor/login` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again - [x] Open a GitHub issue and type some misspellings, make sure they aren't autocorrected - [x] Open an email on `http://mail.google.com/` or inbox.google.com and click on a link. Make sure it works - [x] Verify that `https://mixed-script.badssl.com/` shows up as grey not red (no mixed content scripts are run) ## Brave Rewards/Ads - [x] Verify wallet is auto-created after enabling rewards(either via Panel or Rewards page) - [x] Verify account balance shows correct BAT and USD value - [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [x] Verify grant details are shown in expanded view when a grant is claimed - [x] Verify monthly budget shows correct BAT and USD value - [x] Verify you can exclude a publisher from the auto-contribute table by clicking on the trash bin icon in the auto-contribute table - [x] Verify you can exclude a publisher by using the toggle on the Rewards Panel - [x] Verify you can remove excluded sites via `Restore All` button - [x] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %) - [x] Verify when you click on `Send a tip`, the custom tip banner displays - [x] Verify you can make a one-time tip and they display in tips panel - [x] Verify you can make a recurring tip and they display in tips panel - [x] Verify you can tip a verified publisher - [x] Verify you can tip a verified YouTube creator - [x] Verify tip panel shows a verified check mark for a verified publisher/verified YouTube creator - [x] Verify tip panel shows a message about the unverified publisher - [x] Verify BR panel shows the message about an unverified publisher - [x] Verify you can perform a contribution - [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [x] Verify that disabling Rewards and enabling it again does not lose state - [x] Verify that disabling auto-contribute and enabling it again does not lose state - [x] Verify unchecking `Allow contribution to videos` option doesn't list any YouTube creator in ac list - [x] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [x] Verify you can reset rewards from advance setting. Resetting should delete wallet and bring it back to the pre-optin state - [x] Verify on BR panel `Verify wallet` button loads verify wallet page when balance in >=15 BAT. - [x] Verify user still able to connect to existing Uphold account via panel even when balance is < 15 BAT - [x] Verify you are able to disconnect/re-connect a user wallet when min balance is available - [x] Upgrade from an older version - [x] Verify the wallet balance (if available) is retained - [x] Verify auto-contribute list is not lost after upgrade - [x] Verify tips list is not lost after upgrade - [x] Verify wallet panel transactions list is not lost after upgrade ### Brave Ads - [x] Verify ads is auto-enabled when rewards is enabled for the supported region - [x] Verify ads are only shown when the app is being used - [x] Verify ad notification are shown based on ads per hour setting - [x] Verify ad notifications stack up in notification tray - [x] Verify swipe left/right dismisses the ad notification when shown and is not stored in the notification tray - [x] Verify clicking on an ad notification shows the landing page - [x] Verify `view`,`clicked` and `landed` and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join sync chain by scanning the QR code - [x] Verify you are able to join sync chain using code words - [x] Verify you are able to create a sync chain on the device and add other devices to the chain via QR code/Code words - [x] Verify once sync chain is created, `Categories` option is shown in devices list - [x] Verify only `Bookmarks` is enabled in `Categories` by default - [x] Verify enabling `Sync everything` enables all other switches - [x] Verify existing bookmarks before joining sync chain also gets sync'd to all devices on the sync chain - [x] Verify `Autofill` data is sync'd to all devices when enabled - [x] Verify `History` is sync'd to all devices - [x] Verify `Open tab` form other devices shows up under history below the device name - [x] Verify `Password` is sync'd to all devices - [x] Verify `Settings` sync's site settings on all devices - [x] Verify sync works on an upgrade profile and new bookmarks added post-upgrade sync's across devices on the chain - [x] Verify adding a bookmark on custom tab gets sync'd across all devices in the chain - [x] Verify you are able to create a standalone sync chain with one device - [x] Verify `Remove this device` shows a confirmation alert before removing the sync chain on the device ## Top sites view - [x] Long-press on top sites to get to deletion mode, and delete a top site (note this will stop that site from showing up again on top sites, so you may not want to do this a site you want to keep there) ## Session storage - [x] Verify that tabs restore when closed, including active tab
non_infrastructure
manual test run on android tab for x release installer check that installer is close to the size of the last release check the brave version in about and make sure it is exactly as expected startup verify that brave is only contacting brave com endpoints on first launch using either charles proxy fiddler or wireshark or a similar application visual look make sure thereafter every merge no chrome chromium words appear on brave version no chrome chromium words appear on normal or private tabs no chrome chromium words appear in site settings for location camera microphone augmented reality no chrome chromium icons are shown in normal or private tabs data pre requisite put previous build shortcut on the home screen also have several sites added to home screen from dots menu and then upgrade to new build verify that data from the previous build appears in the updated build as expected bookmarks etc verify that the cookies from the previous build are preserved after upgrade verify shortcut is still available on the home screen after upgrade verify sites added to home screen are still visible and able to be used after upgrade verify sync chain created in the previous version is still retained on upgrade verify settings changes done in the previous version are still retained on upgrade verify ipfs ens unstoppable domain settings is retained after upgrade verify tab group setting doesn t reset upon upgrade verify clear data on exit setting state from previous build is retained upon upgrade custom tabs make sure brave handles links from gmail slack ensure custom tabs work even with sync enabled disabled developer tools verify you can inspect sub links via dev tools clear data verify clear data on exit works as intended settings and bottom bar verify changing default settings are retained and don t cause the browser to crash verify bottom bar buttons home bookmark search tabs work as expected ipfs ens unstoppable domain verify settings for ipfs ens unstoppable domain show up under brave shields privacy settings downloads verify downloading a file works and that all actions on the download item work verify that pdf is downloaded over https at verify that pdf is downloaded over http at bravery settings check that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to check that toggling to blocking and allow ads works as expected verify that clicking through a cert error in works visit and then turn on script blocking nothing should load allow it from the script blocking ui in the url bar and it should work verify that default bravery settings take effect on pages with no site settings verify that party storage results are blank at when party cookies are blocked choose a dns provider from the list in settings privacy use secure dns load and verify your isp s dns resolvers aren t detected and shown only your chosen doh provider should appear run all six test configurations on and confirm results are as expected per each test listed fingerprint tests visit ensure blocked items are listed in shields test that doesn t leak ip address for each option under settings privacy and security webrtc ip handling policy content tests go to and click on the twitter icon on the top right verify that context menus work in the new twitter tab go to and make sure that the password can be saved make sure the saved password is auto populated when you visit the site again open a github issue and type some misspellings make sure they aren t autocorrected open an email on or inbox google com and click on a link make sure it works verify that shows up as grey not red no mixed content scripts are run brave rewards ads verify wallet is auto created after enabling rewards either via panel or rewards page verify account balance shows correct bat and usd value verify actions taken claiming grant tipping auto contribute display in wallet panel verify grant details are shown in expanded view when a grant is claimed verify monthly budget shows correct bat and usd value verify you can exclude a publisher from the auto contribute table by clicking on the trash bin icon in the auto contribute table verify you can exclude a publisher by using the toggle on the rewards panel verify you can remove excluded sites via restore all button verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify when you click on send a tip the custom tip banner displays verify you can make a one time tip and they display in tips panel verify you can make a recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify tip panel shows a verified check mark for a verified publisher verified youtube creator verify tip panel shows a message about the unverified publisher verify br panel shows the message about an unverified publisher verify you can perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators verify that disabling rewards and enabling it again does not lose state verify that disabling auto contribute and enabling it again does not lose state verify unchecking allow contribution to videos option doesn t list any youtube creator in ac list adjust min visit time in settings visit some sites and youtube channels to verify they are added to the table after the specified settings verify you can reset rewards from advance setting resetting should delete wallet and bring it back to the pre optin state verify on br panel verify wallet button loads verify wallet page when balance in bat verify user still able to connect to existing uphold account via panel even when balance is bat verify you are able to disconnect re connect a user wallet when min balance is available upgrade from an older version verify the wallet balance if available is retained verify auto contribute list is not lost after upgrade verify tips list is not lost after upgrade verify wallet panel transactions list is not lost after upgrade brave ads verify ads is auto enabled when rewards is enabled for the supported region verify ads are only shown when the app is being used verify ad notification are shown based on ads per hour setting verify ad notifications stack up in notification tray verify swipe left right dismisses the ad notification when shown and is not stored in the notification tray verify clicking on an ad notification shows the landing page verify view clicked and landed and dismiss states are logged based on the action sync verify you are able to join sync chain by scanning the qr code verify you are able to join sync chain using code words verify you are able to create a sync chain on the device and add other devices to the chain via qr code code words verify once sync chain is created categories option is shown in devices list verify only bookmarks is enabled in categories by default verify enabling sync everything enables all other switches verify existing bookmarks before joining sync chain also gets sync d to all devices on the sync chain verify autofill data is sync d to all devices when enabled verify history is sync d to all devices verify open tab form other devices shows up under history below the device name verify password is sync d to all devices verify settings sync s site settings on all devices verify sync works on an upgrade profile and new bookmarks added post upgrade sync s across devices on the chain verify adding a bookmark on custom tab gets sync d across all devices in the chain verify you are able to create a standalone sync chain with one device verify remove this device shows a confirmation alert before removing the sync chain on the device top sites view long press on top sites to get to deletion mode and delete a top site note this will stop that site from showing up again on top sites so you may not want to do this a site you want to keep there session storage verify that tabs restore when closed including active tab
0
687,370
23,523,872,963
IssuesEvent
2022-08-19 08:55:47
eftafmo/dataviz
https://api.github.com/repos/eftafmo/dataviz
closed
visualisations: use same inactive colour in chart and legend
Priority: Low Type: minor Bug Component: JS
applies to all visualisations with html legend: sectors, mechanisms, beneficiaries the legend should use the same computed colour for inactive items, and drop the opacity / grayscale filter from css. should also take care of #71
1.0
visualisations: use same inactive colour in chart and legend - applies to all visualisations with html legend: sectors, mechanisms, beneficiaries the legend should use the same computed colour for inactive items, and drop the opacity / grayscale filter from css. should also take care of #71
non_infrastructure
visualisations use same inactive colour in chart and legend applies to all visualisations with html legend sectors mechanisms beneficiaries the legend should use the same computed colour for inactive items and drop the opacity grayscale filter from css should also take care of
0
10,499
8,589,881,686
IssuesEvent
2018-11-14 17:31:53
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Add netstandard-RID specific implementations of System.Diagnostics.EventLog & System.Management
area-Infrastructure bug
Unlike other packages in the Windows Compatibility Pack (e.g., `Microsoft.Win32.Registry`), `System.Diagnostics.EventLog` & `System.Management` packages do not provide RID specific implementations for netstandard. Please provide RID specific implementations for netstandard. Please see: https://github.com/dotnet/corefx/issues/32910#issuecomment-431468581
1.0
Add netstandard-RID specific implementations of System.Diagnostics.EventLog & System.Management - Unlike other packages in the Windows Compatibility Pack (e.g., `Microsoft.Win32.Registry`), `System.Diagnostics.EventLog` & `System.Management` packages do not provide RID specific implementations for netstandard. Please provide RID specific implementations for netstandard. Please see: https://github.com/dotnet/corefx/issues/32910#issuecomment-431468581
infrastructure
add netstandard rid specific implementations of system diagnostics eventlog system management unlike other packages in the windows compatibility pack e g microsoft registry system diagnostics eventlog system management packages do not provide rid specific implementations for netstandard please provide rid specific implementations for netstandard please see
1
17,867
12,675,522,602
IssuesEvent
2020-06-19 02:00:16
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Try to dedupe helix content scripts with eng/scripts
area-infrastructure
Ideally we can copy scripts from eng/scripts that are identical
1.0
Try to dedupe helix content scripts with eng/scripts - Ideally we can copy scripts from eng/scripts that are identical
infrastructure
try to dedupe helix content scripts with eng scripts ideally we can copy scripts from eng scripts that are identical
1
27,067
21,096,563,857
IssuesEvent
2022-04-04 10:52:34
epimorphics/hmlr-linked-data
https://api.github.com/repos/epimorphics/hmlr-linked-data
closed
Dev access to SapiNT API
app/PPD infrastructure and tools app/Standard Reports app/UKHPI
We have previously been using ssh tunnels to allow developers to have a `localhost` ssh endpoint that we connect to a locally running Rails application such as PPD. This has been needed because the presentation servers, which host the remote SapiNT instance, are not visible outside of the cluster. @andrew-pickin-epi and @der suggested that an alternative pattern would be to fire up a local Docker container containing the SapiNT instance, talking to the remote Fuseki API.
1.0
Dev access to SapiNT API - We have previously been using ssh tunnels to allow developers to have a `localhost` ssh endpoint that we connect to a locally running Rails application such as PPD. This has been needed because the presentation servers, which host the remote SapiNT instance, are not visible outside of the cluster. @andrew-pickin-epi and @der suggested that an alternative pattern would be to fire up a local Docker container containing the SapiNT instance, talking to the remote Fuseki API.
infrastructure
dev access to sapint api we have previously been using ssh tunnels to allow developers to have a localhost ssh endpoint that we connect to a locally running rails application such as ppd this has been needed because the presentation servers which host the remote sapint instance are not visible outside of the cluster andrew pickin epi and der suggested that an alternative pattern would be to fire up a local docker container containing the sapint instance talking to the remote fuseki api
1
242,934
18,674,171,213
IssuesEvent
2021-10-31 09:04:08
minicli/docs
https://api.github.com/repos/minicli/docs
closed
Improve docs organization
documentation
The current documentation has only a few pages and they aren't well organized. We could probably use a better structure with sub sections.
1.0
Improve docs organization - The current documentation has only a few pages and they aren't well organized. We could probably use a better structure with sub sections.
non_infrastructure
improve docs organization the current documentation has only a few pages and they aren t well organized we could probably use a better structure with sub sections
0
71,904
18,923,046,729
IssuesEvent
2021-11-17 05:43:44
ARM-software/armnn
https://api.github.com/repos/ARM-software/armnn
closed
Compile error for TFlite Delegate
Documentation issue Build issue
aim at: **Build the TfLite Delegate (Stand-Alone)** ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md env: **1) ubuntu18.04(x86) 2) tensorflow-2.3.1 -build success 3) flatbuffers-1.12.0-build arm64 success 4) compute library_21.02 -build success** `scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0` **5)build armnn 21.02 success** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0 make ``` **6) build tflte delegate** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn make ``` /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all look at dirctory,libtensorflow_lite_all.so is exist: ``` ls ../../../tensorflow-bazel-bin libtensorflow_lite_all.so ``` please give me some advice to slove it,thanks。
1.0
Compile error for TFlite Delegate - aim at: **Build the TfLite Delegate (Stand-Alone)** ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md env: **1) ubuntu18.04(x86) 2) tensorflow-2.3.1 -build success 3) flatbuffers-1.12.0-build arm64 success 4) compute library_21.02 -build success** `scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0` **5)build armnn 21.02 success** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0 make ``` **6) build tflte delegate** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn make ``` /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all look at dirctory,libtensorflow_lite_all.so is exist: ``` ls ../../../tensorflow-bazel-bin libtensorflow_lite_all.so ``` please give me some advice to slove it,thanks。
non_infrastructure
compile error for tflite delegate aim at build the tflite delegate stand alone ref from env tensorflow build success flatbuffers build success compute library build success scons arch neon opencl embed kernels extra cxx flags fpic benchmark tests validation tests internal only build armnn success cxx linux gnu g cc linux gnu gcc cmake darmcompute root basedir computelibrary darmcomputeneon dbuild unit tests make build tflte delegate cxx linux gnu g cc linux gnu gcc cmake dtensorflow lib dir basedir tensorflow bazel bin dtensorflow root basedir tensorflow dtflite lib root basedir tensorflow bazel bin dflatbuffers root basedir flatbuffers install darmnn dir basedir armnn build darmnn source dir basedir armnn make usr lib gcc cross linux gnu linux gnu bin ld skipping incompatible home delegate tensorflow bazel bin libtensorflow lite all so when searching for ltensorflow lite all usr lib gcc cross linux gnu linux gnu bin ld cannot find ltensorflow lite all look at dirctory,libtensorflow lite all so is exist: ls tensorflow bazel bin libtensorflow lite all so please give me some advice to slove it,thanks。
0
166,829
26,414,142,262
IssuesEvent
2023-01-13 14:39:45
cowprotocol/cowswap
https://api.github.com/repos/cowprotocol/cowswap
closed
Scrollbar is displayed under the modal
app:CowSwap Bug Design Low
Related to #1785 1. Connect to a wallet 2. Open Account modal and scroll down its content 3. Check the right top and bottom corner **AR**: Scrollbar is displayed under the modal ![scrollbar.png](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/339a67d3-5f87-4eae-be37-daf53d021703)![scrollbar under.png](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/37cc5f21-958e-437c-8712-8a90108fb723) **ER**: Scrollbar does not exceed the height of a modal
1.0
Scrollbar is displayed under the modal - Related to #1785 1. Connect to a wallet 2. Open Account modal and scroll down its content 3. Check the right top and bottom corner **AR**: Scrollbar is displayed under the modal ![scrollbar.png](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/339a67d3-5f87-4eae-be37-daf53d021703)![scrollbar under.png](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/37cc5f21-958e-437c-8712-8a90108fb723) **ER**: Scrollbar does not exceed the height of a modal
non_infrastructure
scrollbar is displayed under the modal related to connect to a wallet open account modal and scroll down its content check the right top and bottom corner ar scrollbar is displayed under the modal er scrollbar does not exceed the height of a modal
0
188,009
14,436,126,350
IssuesEvent
2020-12-07 09:42:48
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
xdvproject/ldap: dn_test.go; 22 LoC
fresh small test
Found a possible issue in [xdvproject/ldap](https://www.github.com/xdvproject/ldap) at [dn_test.go](https://github.com/xdvproject/ldap/blob/92584199f797edcb0f3777d6ad6cd94706981cf5/dn_test.go#L52-L73) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to answer at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/xdvproject/ldap/blob/92584199f797edcb0f3777d6ad6cd94706981cf5/dn_test.go#L52-L73) <details> <summary>Click here to show the 22 line(s) of Go which triggered the analyzer.</summary> ```go for test, answer := range testcases { dn, err := ldap.ParseDN(test) if err != nil { t.Errorf(err.Error()) continue } if !reflect.DeepEqual(dn, &answer) { t.Errorf("Parsed DN %s is not equal to the expected structure", test) t.Logf("Expected:") for _, rdn := range answer.RDNs { for _, attribs := range rdn.Attributes { t.Logf("#%v\n", attribs) } } t.Logf("Actual:") for _, rdn := range dn.RDNs { for _, attribs := range rdn.Attributes { t.Logf("#%v\n", attribs) } } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 92584199f797edcb0f3777d6ad6cd94706981cf5
1.0
xdvproject/ldap: dn_test.go; 22 LoC - Found a possible issue in [xdvproject/ldap](https://www.github.com/xdvproject/ldap) at [dn_test.go](https://github.com/xdvproject/ldap/blob/92584199f797edcb0f3777d6ad6cd94706981cf5/dn_test.go#L52-L73) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to answer at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/xdvproject/ldap/blob/92584199f797edcb0f3777d6ad6cd94706981cf5/dn_test.go#L52-L73) <details> <summary>Click here to show the 22 line(s) of Go which triggered the analyzer.</summary> ```go for test, answer := range testcases { dn, err := ldap.ParseDN(test) if err != nil { t.Errorf(err.Error()) continue } if !reflect.DeepEqual(dn, &answer) { t.Errorf("Parsed DN %s is not equal to the expected structure", test) t.Logf("Expected:") for _, rdn := range answer.RDNs { for _, attribs := range rdn.Attributes { t.Logf("#%v\n", attribs) } } t.Logf("Actual:") for _, rdn := range dn.RDNs { for _, attribs := range rdn.Attributes { t.Logf("#%v\n", attribs) } } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 92584199f797edcb0f3777d6ad6cd94706981cf5
non_infrastructure
xdvproject ldap dn test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to answer at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test answer range testcases dn err ldap parsedn test if err nil t errorf err error continue if reflect deepequal dn answer t errorf parsed dn s is not equal to the expected structure test t logf expected for rdn range answer rdns for attribs range rdn attributes t logf v n attribs t logf actual for rdn range dn rdns for attribs range rdn attributes t logf v n attribs leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
3,041
4,028,606,916
IssuesEvent
2016-05-18 07:17:03
camptocamp/ngeo
https://api.github.com/repos/camptocamp/ngeo
closed
make transifex-get not working
Infrastructure Ready
``` $ make transifex-get .build/locale/gmf.pot make: .build/locale/gmf.pot: Command not found Makefile:687: recipe for target 'transifex-get' failed make: *** [transifex-get] Error 127 ```
1.0
make transifex-get not working - ``` $ make transifex-get .build/locale/gmf.pot make: .build/locale/gmf.pot: Command not found Makefile:687: recipe for target 'transifex-get' failed make: *** [transifex-get] Error 127 ```
infrastructure
make transifex get not working make transifex get build locale gmf pot make build locale gmf pot command not found makefile recipe for target transifex get failed make error
1
20,125
13,676,576,214
IssuesEvent
2020-09-29 14:03:31
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
opened
Karma tests for various screen sizes
Pod: Pea (Dashboard & Templates) Pod: Prometheus (Workspace) Pod: WP & Infra Type: Infrastructure
<!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Task Description Splitting this out from https://github.com/google/web-stories-wp/pull/2157#issuecomment-636995336 so it can be tracked appropriately. It would be great if we could run the Karma tests multiple times, for different screen sizes. Something like `npm run test:karma -- --viewport=w1:h1; npm run test:karma -- --viewport=w2:h2` Some API needs to be defined to enable writing different tests targeting specific viewports.
1.0
Karma tests for various screen sizes - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Task Description Splitting this out from https://github.com/google/web-stories-wp/pull/2157#issuecomment-636995336 so it can be tracked appropriately. It would be great if we could run the Karma tests multiple times, for different screen sizes. Something like `npm run test:karma -- --viewport=w1:h1; npm run test:karma -- --viewport=w2:h2` Some API needs to be defined to enable writing different tests targeting specific viewports.
infrastructure
karma tests for various screen sizes task description splitting this out from so it can be tracked appropriately it would be great if we could run the karma tests multiple times for different screen sizes something like npm run test karma viewport npm run test karma viewport some api needs to be defined to enable writing different tests targeting specific viewports
1
34,906
30,572,042,585
IssuesEvent
2023-07-20 23:38:23
civiform/civiform
https://api.github.com/repos/civiform/civiform
closed
Add APPLICANT_REGISTER_URI to deploy template and TF deploy system
bug infrastructure dow-swe deploy
`APPLICANT_REGISTER_URI` is not passed through the deployment system. With it missing, the register link instead redirects applicants to the login action instead of the registration page.
1.0
Add APPLICANT_REGISTER_URI to deploy template and TF deploy system - `APPLICANT_REGISTER_URI` is not passed through the deployment system. With it missing, the register link instead redirects applicants to the login action instead of the registration page.
infrastructure
add applicant register uri to deploy template and tf deploy system applicant register uri is not passed through the deployment system with it missing the register link instead redirects applicants to the login action instead of the registration page
1
30,262
24,708,676,702
IssuesEvent
2022-10-19 21:34:00
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
The 3.1 and 5.0 SDKs do not support musl/Alpine Linux anymore
area-Infrastructure-libraries untriaged
### Describe the bug I am building an app for Alpine 3.15 and/or 3.16, but those versions of Alpine can no longer load the .so files from the linux-musl-x64 folders. I had a look at the runtime.json on those installs and I notice the version of alpine in the json is only 3.14. ### To Reproduce Just try use SkiaSharp - or any other NuGet that has a musl-specific binary. Se also these issues: - https://github.com/mono/SkiaSharp/issues/2169 - https://github.com/mono/SkiaSharp/issues/2215 I can work around this by adding/editing the json file to include alpine.3.15 and alpine.3.16. ### Exceptions (if any) ``` System.DllNotFoundException: Unable to load shared library 'libSkiaSharp' or one of its dependencies. In order to help diagnose loading problems, consider setting the LD_DEBUG environment variable: Error loading shared library liblibSkiaSharp: No such file or directory ``` ### Further technical details - Using the dotnet docker image which was Alpine 3.16 and had SDK 3.1.28 installed. - I also see the Alpine version in the json on GitHub: https://github.com/dotnet/corefx/blob/v3.1.28/pkg/Microsoft.NETCore.Platforms/runtime.json#L80
1.0
The 3.1 and 5.0 SDKs do not support musl/Alpine Linux anymore - ### Describe the bug I am building an app for Alpine 3.15 and/or 3.16, but those versions of Alpine can no longer load the .so files from the linux-musl-x64 folders. I had a look at the runtime.json on those installs and I notice the version of alpine in the json is only 3.14. ### To Reproduce Just try use SkiaSharp - or any other NuGet that has a musl-specific binary. Se also these issues: - https://github.com/mono/SkiaSharp/issues/2169 - https://github.com/mono/SkiaSharp/issues/2215 I can work around this by adding/editing the json file to include alpine.3.15 and alpine.3.16. ### Exceptions (if any) ``` System.DllNotFoundException: Unable to load shared library 'libSkiaSharp' or one of its dependencies. In order to help diagnose loading problems, consider setting the LD_DEBUG environment variable: Error loading shared library liblibSkiaSharp: No such file or directory ``` ### Further technical details - Using the dotnet docker image which was Alpine 3.16 and had SDK 3.1.28 installed. - I also see the Alpine version in the json on GitHub: https://github.com/dotnet/corefx/blob/v3.1.28/pkg/Microsoft.NETCore.Platforms/runtime.json#L80
infrastructure
the and sdks do not support musl alpine linux anymore describe the bug i am building an app for alpine and or but those versions of alpine can no longer load the so files from the linux musl folders i had a look at the runtime json on those installs and i notice the version of alpine in the json is only to reproduce just try use skiasharp or any other nuget that has a musl specific binary se also these issues i can work around this by adding editing the json file to include alpine and alpine exceptions if any system dllnotfoundexception unable to load shared library libskiasharp or one of its dependencies in order to help diagnose loading problems consider setting the ld debug environment variable error loading shared library liblibskiasharp no such file or directory further technical details using the dotnet docker image which was alpine and had sdk installed i also see the alpine version in the json on github
1
19,708
14,480,258,663
IssuesEvent
2020-12-10 10:55:25
virtualsatellite/VirtualSatellite4-Core
https://api.github.com/repos/virtualsatellite/VirtualSatellite4-Core
closed
Create a navigator filter for CompossedPropertyInstances
comfort/usability ui
Compossed porperty instance bloat the system model presentation without providong useful content... We should add some kind of filter mechanism to hide them if not needed...
True
Create a navigator filter for CompossedPropertyInstances - Compossed porperty instance bloat the system model presentation without providong useful content... We should add some kind of filter mechanism to hide them if not needed...
non_infrastructure
create a navigator filter for compossedpropertyinstances compossed porperty instance bloat the system model presentation without providong useful content we should add some kind of filter mechanism to hide them if not needed
0
61,158
8,490,676,451
IssuesEvent
2018-10-27 03:47:53
saltstack/salt
https://api.github.com/repos/saltstack/salt
closed
GPG Renderer parse flag
Bug Core Documentation Medium Severity P4 Renderers stale
### Description of Issue/Question In the GPG renderer documentation it mentions that in order to render the file you need to add `#!yaml|gpg` however doing so implicitly disables the jinja renderer. This is probably the desired functionality in general but it is unclear in the documentation (and only seems to occur if the render statement is at the top of the file). Either the functionality should be updated to still allow the default renderers or the gpg renderer documentation should be updated. ### Setup ``` #!yaml|gpg {% if grains['test'] = true %} some_secret: | -----BEGIN PGP MESSAGE----- Version: GnuPG v2.0.22 (GNU/Linux) <BLAH> -----END PGP MESSAGE----- {% endif %} ``` ### Steps to Reproduce Issue ``` 2017-05-08 15:46:13,146 [salt.pillar ][CRITICAL][23728] Rendering SLS 'test_gpg' failed, render error: found character '%' that cannot start any token; line 2 --- #!yaml|gpg {% if grains['test'] == 'test' %} <====================== some_secret: | -----BEGIN PGP MESSAGE----- Version: GnuPG v2.0.22 (GNU/Linux) <BLAH> -----END PGP MESSAGE----- [...] ``` #### Workaround: This can be fixed in two situations: 1. If the `#!yaml|gpg` is the second line to another comment 2. If you add `#!jinja|yaml|gpg` ### Versions Report ``` Salt Version: Salt: 2016.11.4 Dependency Versions: cffi: 1.6.0 cherrypy: Not Installed dateutil: 1.5 docker-py: Not Installed gitdb: Not Installed gitpython: Not Installed ioflo: Not Installed Jinja2: 2.7.2 libgit2: 0.24.6 libnacl: Not Installed M2Crypto: 0.21.1 Mako: Not Installed msgpack-pure: Not Installed msgpack-python: 0.4.8 mysql-python: Not Installed pycparser: 2.14 pycrypto: 2.6.1 pycryptodome: 3.4.3 pygit2: 0.24.2 Python: 2.7.5 (default, Aug 2 2016, 04:20:16) python-gnupg: Not Installed PyYAML: 3.11 PyZMQ: 15.3.0 RAET: Not Installed smmap: Not Installed timelib: Not Installed Tornado: 4.2.1 ZMQ: 4.1.4 System Versions: dist: redhat 7.3 Maipo machine: x86_64 release: 3.10.0-327.10.1.el7.x86_64 system: Linux version: Red Hat Enterprise Linux Server 7.3 Maipo ```
1.0
GPG Renderer parse flag - ### Description of Issue/Question In the GPG renderer documentation it mentions that in order to render the file you need to add `#!yaml|gpg` however doing so implicitly disables the jinja renderer. This is probably the desired functionality in general but it is unclear in the documentation (and only seems to occur if the render statement is at the top of the file). Either the functionality should be updated to still allow the default renderers or the gpg renderer documentation should be updated. ### Setup ``` #!yaml|gpg {% if grains['test'] = true %} some_secret: | -----BEGIN PGP MESSAGE----- Version: GnuPG v2.0.22 (GNU/Linux) <BLAH> -----END PGP MESSAGE----- {% endif %} ``` ### Steps to Reproduce Issue ``` 2017-05-08 15:46:13,146 [salt.pillar ][CRITICAL][23728] Rendering SLS 'test_gpg' failed, render error: found character '%' that cannot start any token; line 2 --- #!yaml|gpg {% if grains['test'] == 'test' %} <====================== some_secret: | -----BEGIN PGP MESSAGE----- Version: GnuPG v2.0.22 (GNU/Linux) <BLAH> -----END PGP MESSAGE----- [...] ``` #### Workaround: This can be fixed in two situations: 1. If the `#!yaml|gpg` is the second line to another comment 2. If you add `#!jinja|yaml|gpg` ### Versions Report ``` Salt Version: Salt: 2016.11.4 Dependency Versions: cffi: 1.6.0 cherrypy: Not Installed dateutil: 1.5 docker-py: Not Installed gitdb: Not Installed gitpython: Not Installed ioflo: Not Installed Jinja2: 2.7.2 libgit2: 0.24.6 libnacl: Not Installed M2Crypto: 0.21.1 Mako: Not Installed msgpack-pure: Not Installed msgpack-python: 0.4.8 mysql-python: Not Installed pycparser: 2.14 pycrypto: 2.6.1 pycryptodome: 3.4.3 pygit2: 0.24.2 Python: 2.7.5 (default, Aug 2 2016, 04:20:16) python-gnupg: Not Installed PyYAML: 3.11 PyZMQ: 15.3.0 RAET: Not Installed smmap: Not Installed timelib: Not Installed Tornado: 4.2.1 ZMQ: 4.1.4 System Versions: dist: redhat 7.3 Maipo machine: x86_64 release: 3.10.0-327.10.1.el7.x86_64 system: Linux version: Red Hat Enterprise Linux Server 7.3 Maipo ```
non_infrastructure
gpg renderer parse flag description of issue question in the gpg renderer documentation it mentions that in order to render the file you need to add yaml gpg however doing so implicitly disables the jinja renderer this is probably the desired functionality in general but it is unclear in the documentation and only seems to occur if the render statement is at the top of the file either the functionality should be updated to still allow the default renderers or the gpg renderer documentation should be updated setup yaml gpg if grains true some secret begin pgp message version gnupg gnu linux end pgp message endif steps to reproduce issue rendering sls test gpg failed render error found character that cannot start any token line yaml gpg if grains test some secret begin pgp message version gnupg gnu linux end pgp message workaround this can be fixed in two situations if the yaml gpg is the second line to another comment if you add jinja yaml gpg versions report salt version salt dependency versions cffi cherrypy not installed dateutil docker py not installed gitdb not installed gitpython not installed ioflo not installed libnacl not installed mako not installed msgpack pure not installed msgpack python mysql python not installed pycparser pycrypto pycryptodome python default aug python gnupg not installed pyyaml pyzmq raet not installed smmap not installed timelib not installed tornado zmq system versions dist redhat maipo machine release system linux version red hat enterprise linux server maipo
0
293,833
22,091,725,473
IssuesEvent
2022-06-01 06:41:02
KorBasilion/OSS-Basic-Project
https://api.github.com/repos/KorBasilion/OSS-Basic-Project
closed
new window 여는 함수
documentation
`def openNewWindow(info): newWindow = Toplevel() newWindow.title("Help") newWindow.grid Label(newWindow, text = info, fg = "red").pack()` the "info" parameter takes a text that you want the new window to have. For example: `text2 = "We are CBNU students" Us = Button(settingFrame, image=test, text="About Us", width=388, height=50, compound="c", fg="dark green", command = lambda : openNewWindow(text2))` The result: when you click Us button, the new window opens with a text defined in a text2 variable.
1.0
new window 여는 함수 - `def openNewWindow(info): newWindow = Toplevel() newWindow.title("Help") newWindow.grid Label(newWindow, text = info, fg = "red").pack()` the "info" parameter takes a text that you want the new window to have. For example: `text2 = "We are CBNU students" Us = Button(settingFrame, image=test, text="About Us", width=388, height=50, compound="c", fg="dark green", command = lambda : openNewWindow(text2))` The result: when you click Us button, the new window opens with a text defined in a text2 variable.
non_infrastructure
new window 여는 함수 def opennewwindow info newwindow toplevel newwindow title help newwindow grid label newwindow text info fg red pack the info parameter takes a text that you want the new window to have for example we are cbnu students us button settingframe image test text about us width height compound c fg dark green command lambda opennewwindow the result when you click us button the new window opens with a text defined in a variable
0
383,273
11,353,732,452
IssuesEvent
2020-01-24 16:07:25
cBioPortal/cbioportal
https://api.github.com/repos/cBioPortal/cbioportal
opened
GENIE public portal only show MUT vs CNA chart by default
bug priority study page
We (@tmazor and I) noticed that only MUT vs CNA chart was shown by default for the [genie study](https://genie.cbioportal.org/study/summary?id=genie_public). Reset charts will get charts back. @kalletlak any thoughts on this? cc'ing @n1zea144 @tmazor
1.0
GENIE public portal only show MUT vs CNA chart by default - We (@tmazor and I) noticed that only MUT vs CNA chart was shown by default for the [genie study](https://genie.cbioportal.org/study/summary?id=genie_public). Reset charts will get charts back. @kalletlak any thoughts on this? cc'ing @n1zea144 @tmazor
non_infrastructure
genie public portal only show mut vs cna chart by default we tmazor and i noticed that only mut vs cna chart was shown by default for the reset charts will get charts back kalletlak any thoughts on this cc ing tmazor
0
393,074
11,609,802,598
IssuesEvent
2020-02-26 01:07:43
frc3452/2020_FRC
https://api.github.com/repos/frc3452/2020_FRC
opened
Polish up Teleop code
priority_1
Polish up the Teleop code and make sure it's ready for the drive team. - Change the button layout to something more intuitive - Remove the three intake speeds, they're not needed - Change the speed of the driving and the turning until it feels more intuitive - Anything else the drive team requests
1.0
Polish up Teleop code - Polish up the Teleop code and make sure it's ready for the drive team. - Change the button layout to something more intuitive - Remove the three intake speeds, they're not needed - Change the speed of the driving and the turning until it feels more intuitive - Anything else the drive team requests
non_infrastructure
polish up teleop code polish up the teleop code and make sure it s ready for the drive team change the button layout to something more intuitive remove the three intake speeds they re not needed change the speed of the driving and the turning until it feels more intuitive anything else the drive team requests
0
28,107
22,973,171,993
IssuesEvent
2022-07-20 06:19:07
bcgov/foi-flow
https://api.github.com/repos/bcgov/foi-flow
opened
Analysis and Action on PROD's variable data storage/Back up storage Optimization - KINDLING(File Store) , DB( SPHINX)
devops Task Data Infrastructure
Title of ticket: #### Description The purpose of this task is to analyze back up procedure on KINDLING.IDIR.BCGOV (File Store) and how data related to AXIS is mapped with FileStore. So that, we can better handle purge process of older data and also, acquire knowledge on data restoration. #### Dependencies Are there any dependencies? #### DOD - [ ] Analysis on back up procedure on KINDLING.IDIR.BCGOV - [ ] Decide on the option to purge old files (<2019). - [ ] Analyze and decide the option to purge related data on PROD DB. - [ ] Analyze the option to set up File store back up to OSS / S3 - [ ] Purge the file(s), folder(s) as per the decision made - [ ] Purge data on DB - review the notes from TEST PURGING
1.0
Analysis and Action on PROD's variable data storage/Back up storage Optimization - KINDLING(File Store) , DB( SPHINX) - Title of ticket: #### Description The purpose of this task is to analyze back up procedure on KINDLING.IDIR.BCGOV (File Store) and how data related to AXIS is mapped with FileStore. So that, we can better handle purge process of older data and also, acquire knowledge on data restoration. #### Dependencies Are there any dependencies? #### DOD - [ ] Analysis on back up procedure on KINDLING.IDIR.BCGOV - [ ] Decide on the option to purge old files (<2019). - [ ] Analyze and decide the option to purge related data on PROD DB. - [ ] Analyze the option to set up File store back up to OSS / S3 - [ ] Purge the file(s), folder(s) as per the decision made - [ ] Purge data on DB - review the notes from TEST PURGING
infrastructure
analysis and action on prod s variable data storage back up storage optimization kindling file store db sphinx title of ticket description the purpose of this task is to analyze back up procedure on kindling idir bcgov file store and how data related to axis is mapped with filestore so that we can better handle purge process of older data and also acquire knowledge on data restoration dependencies are there any dependencies dod analysis on back up procedure on kindling idir bcgov decide on the option to purge old files analyze and decide the option to purge related data on prod db analyze the option to set up file store back up to oss purge the file s folder s as per the decision made purge data on db review the notes from test purging
1
84,954
24,474,306,456
IssuesEvent
2022-10-08 01:43:37
arkamnite/llvm-project
https://api.github.com/repos/arkamnite/llvm-project
opened
(bug): GameBoyRegisterInfo is an abstract class
register set build bug
## `GameBoyRegisterInfo` is an abstract class
1.0
(bug): GameBoyRegisterInfo is an abstract class - ## `GameBoyRegisterInfo` is an abstract class
non_infrastructure
bug gameboyregisterinfo is an abstract class gameboyregisterinfo is an abstract class
0
20,442
13,923,056,730
IssuesEvent
2020-10-21 14:00:05
hackcambridge/hc2020-prototype
https://api.github.com/repos/hackcambridge/hc2020-prototype
closed
Speed up CodeDeploy pipeline
enhancement infrastructure long term medium priority
Currently 7 minutes. 3 would be ideal to speed up dev testing on canary.
1.0
Speed up CodeDeploy pipeline - Currently 7 minutes. 3 would be ideal to speed up dev testing on canary.
infrastructure
speed up codedeploy pipeline currently minutes would be ideal to speed up dev testing on canary
1
32,506
26,746,701,460
IssuesEvent
2023-01-30 16:27:21
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
Explore allowing an administrator to enable/disable ODH components from UI
kind/enhancement infrastructure priority/normal
With access to the `kfdef` file, we could allow authorized users to update the installed components of Open Data Hub. This would require authentication through "Login with OpenShift" and verify they have access. You could allow users to customize kfdef fields with a forms wizard. Requirements: - Allow users to Login with OpenShift - Allow components to be correctly: - installed - updated - uninstalled ![Screenshot from 2020-08-31 13-31-19](https://user-images.githubusercontent.com/1448375/91748639-4a245580-eb8e-11ea-9739-c8c233bb7baf.png) ![Screenshot from 2020-08-31 13-31-23](https://user-images.githubusercontent.com/1448375/91748635-485a9200-eb8e-11ea-9f8e-de0660e25031.png) ![Screenshot from 2020-08-31 13-30-58](https://user-images.githubusercontent.com/1448375/91748608-3c6ed000-eb8e-11ea-870a-343ea9c9114c.png)
1.0
Explore allowing an administrator to enable/disable ODH components from UI - With access to the `kfdef` file, we could allow authorized users to update the installed components of Open Data Hub. This would require authentication through "Login with OpenShift" and verify they have access. You could allow users to customize kfdef fields with a forms wizard. Requirements: - Allow users to Login with OpenShift - Allow components to be correctly: - installed - updated - uninstalled ![Screenshot from 2020-08-31 13-31-19](https://user-images.githubusercontent.com/1448375/91748639-4a245580-eb8e-11ea-9739-c8c233bb7baf.png) ![Screenshot from 2020-08-31 13-31-23](https://user-images.githubusercontent.com/1448375/91748635-485a9200-eb8e-11ea-9f8e-de0660e25031.png) ![Screenshot from 2020-08-31 13-30-58](https://user-images.githubusercontent.com/1448375/91748608-3c6ed000-eb8e-11ea-870a-343ea9c9114c.png)
infrastructure
explore allowing an administrator to enable disable odh components from ui with access to the kfdef file we could allow authorized users to update the installed components of open data hub this would require authentication through login with openshift and verify they have access you could allow users to customize kfdef fields with a forms wizard requirements allow users to login with openshift allow components to be correctly installed updated uninstalled
1
26,166
6,755,440,057
IssuesEvent
2017-10-24 00:31:44
jascam/CodePlexFoo
https://api.github.com/repos/jascam/CodePlexFoo
closed
Create Example: CSCodeDOM
CodePlexMigrationInitiated enhancement impact: Medium
The CSCodeDOM project demonstrates how to use the .NET CodeDOM mechanism to enable dynamic souce code generation and compilation at runtime. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '3259' Vote count: '1'
1.0
Create Example: CSCodeDOM - The CSCodeDOM project demonstrates how to use the .NET CodeDOM mechanism to enable dynamic souce code generation and compilation at runtime. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '3259' Vote count: '1'
non_infrastructure
create example cscodedom the cscodedom project demonstrates how to use the net codedom mechanism to enable dynamic souce code generation and compilation at runtime migrated codeplex work item details codeplex work item id vote count
0
25,766
6,712,739,887
IssuesEvent
2017-10-13 10:32:28
eclipse/che
https://api.github.com/repos/eclipse/che
closed
Validation does not work in the "Build stack from recipe" form
kind/bug severity/P2 sprint/next-sprint status/code-review team/plugin
**Reproduction Steps:** - Go to dashboard - Click "Create Workspace" button - Click "Add Stack button" - Click "Yes" button - Wait until "Build stack from recipe" form will be opened - Type not valid text (for example "@##$$$#@#$" ) - Pay attention that error message is not displayed - Click "Ok" button - Pay attention that form is still open and error message is not shown **OS and version:** CHE 5.16.0 ; Chrome 59.0 ; Fedora 25 **Attachment:** ![second](https://user-images.githubusercontent.com/22724619/28671951-004f61c0-72e7-11e7-905e-d43d553d72c7.gif)
1.0
Validation does not work in the "Build stack from recipe" form - **Reproduction Steps:** - Go to dashboard - Click "Create Workspace" button - Click "Add Stack button" - Click "Yes" button - Wait until "Build stack from recipe" form will be opened - Type not valid text (for example "@##$$$#@#$" ) - Pay attention that error message is not displayed - Click "Ok" button - Pay attention that form is still open and error message is not shown **OS and version:** CHE 5.16.0 ; Chrome 59.0 ; Fedora 25 **Attachment:** ![second](https://user-images.githubusercontent.com/22724619/28671951-004f61c0-72e7-11e7-905e-d43d553d72c7.gif)
non_infrastructure
validation does not work in the build stack from recipe form reproduction steps go to dashboard click create workspace button click add stack button click yes button wait until build stack from recipe form will be opened type not valid text for example pay attention that error message is not displayed click ok button pay attention that form is still open and error message is not shown os and version che chrome fedora attachment
0
20,755
14,138,693,075
IssuesEvent
2020-11-10 08:48:49
fluencelabs/fluence
https://api.github.com/repos/fluencelabs/fluence
closed
Monitor status of the netdata agents
~infrastructure
Currently, node metrics are collected by small netdata agents. Their only job is to collect and send metrics to the central netdata. However, they still exit for different reasons. As it's hard to understand the whole picture of their behaviour, it would be useful to monitor them through Nagios. Thoughts: there may be some NRPE-like solution that checks that container named `netdata` is running on all nodes.
1.0
Monitor status of the netdata agents - Currently, node metrics are collected by small netdata agents. Their only job is to collect and send metrics to the central netdata. However, they still exit for different reasons. As it's hard to understand the whole picture of their behaviour, it would be useful to monitor them through Nagios. Thoughts: there may be some NRPE-like solution that checks that container named `netdata` is running on all nodes.
infrastructure
monitor status of the netdata agents currently node metrics are collected by small netdata agents their only job is to collect and send metrics to the central netdata however they still exit for different reasons as it s hard to understand the whole picture of their behaviour it would be useful to monitor them through nagios thoughts there may be some nrpe like solution that checks that container named netdata is running on all nodes
1
18,859
13,146,217,645
IssuesEvent
2020-08-08 08:30:57
vuejs/vetur
https://api.github.com/repos/vuejs/vetur
closed
Test failure does not trigger CI failure
bug infrastructure
<!-- Check those before opening an issue --> - [x] I have searched through existing issues - [x] I have read through [docs](https://vuejs.github.io/vetur) - [x] I have read [FAQ](https://github.com/vuejs/vetur/blob/master/docs/FAQ.md) ## Info - Platform: Any - Vetur version: master - VS Code version: 1.47.3 ## Problem <!-- Include error message from Panel -> Output -> Vue Language Server --> <!-- With screenshot / gif if possible --> https://github.com/vuejs/vetur/runs/961144612#step:10:610 ## Reproducible Case <!-- Please provide clear steps for reproducing the problem. Otherwise we might close your issue. - Generate a project from https://github.com/octref/veturpack/generate - Make a minimal code change to demonstrate your problem - Push your commit - Include link to your changes in the issue --> In master branch.
1.0
Test failure does not trigger CI failure - <!-- Check those before opening an issue --> - [x] I have searched through existing issues - [x] I have read through [docs](https://vuejs.github.io/vetur) - [x] I have read [FAQ](https://github.com/vuejs/vetur/blob/master/docs/FAQ.md) ## Info - Platform: Any - Vetur version: master - VS Code version: 1.47.3 ## Problem <!-- Include error message from Panel -> Output -> Vue Language Server --> <!-- With screenshot / gif if possible --> https://github.com/vuejs/vetur/runs/961144612#step:10:610 ## Reproducible Case <!-- Please provide clear steps for reproducing the problem. Otherwise we might close your issue. - Generate a project from https://github.com/octref/veturpack/generate - Make a minimal code change to demonstrate your problem - Push your commit - Include link to your changes in the issue --> In master branch.
infrastructure
test failure does not trigger ci failure i have searched through existing issues i have read through i have read info platform any vetur version master vs code version problem output vue language server reproducible case please provide clear steps for reproducing the problem otherwise we might close your issue generate a project from make a minimal code change to demonstrate your problem push your commit include link to your changes in the issue in master branch
1
531,549
15,500,113,612
IssuesEvent
2021-03-11 08:54:27
teamforus/forus
https://api.github.com/repos/teamforus/forus
closed
Let sponsor add/update actions: Fund configuration
Priority: Must have
## Main asssignee: @dev-rminds CR: https://github.com/teamforus/general/issues/699 Figma: https://www.figma.com/file/zhX7Ra83QmhftuT93U06E6/699%3A-Let-sponsor-add%2Fupdate-actions?node-id=176%3A41175&viewport=-183%2C-109%2C0.5033360123634338 ## Context/goal: Implement the fund configuration that implies the possibility to allow the sponsor to add actions on behalf of the provider. This implies: - Checkbox on the fund creation / edit page. ![Screenshot 2021-02-05 at 12 13 01](https://user-images.githubusercontent.com/33207036/107026785-8433fc00-67ab-11eb-9ad3-4ff3bba18acb.png) - Add this configuration to the ENV or in the database to the fund - Use this configuration to allow or not adding products on behalf of the provider (all providers approved for this fund agree automatically with this configuration).
1.0
Let sponsor add/update actions: Fund configuration - ## Main asssignee: @dev-rminds CR: https://github.com/teamforus/general/issues/699 Figma: https://www.figma.com/file/zhX7Ra83QmhftuT93U06E6/699%3A-Let-sponsor-add%2Fupdate-actions?node-id=176%3A41175&viewport=-183%2C-109%2C0.5033360123634338 ## Context/goal: Implement the fund configuration that implies the possibility to allow the sponsor to add actions on behalf of the provider. This implies: - Checkbox on the fund creation / edit page. ![Screenshot 2021-02-05 at 12 13 01](https://user-images.githubusercontent.com/33207036/107026785-8433fc00-67ab-11eb-9ad3-4ff3bba18acb.png) - Add this configuration to the ENV or in the database to the fund - Use this configuration to allow or not adding products on behalf of the provider (all providers approved for this fund agree automatically with this configuration).
non_infrastructure
let sponsor add update actions fund configuration main asssignee dev rminds cr figma context goal implement the fund configuration that implies the possibility to allow the sponsor to add actions on behalf of the provider this implies checkbox on the fund creation edit page add this configuration to the env or in the database to the fund use this configuration to allow or not adding products on behalf of the provider all providers approved for this fund agree automatically with this configuration
0
290,018
32,029,807,725
IssuesEvent
2023-09-22 11:29:46
dreamboy9/mongo
https://api.github.com/repos/dreamboy9/mongo
closed
CVE-2022-45406 (Critical) detected in mongor5.0.0-rc5 - autoclosed
Mend: dependency security vulnerability
## CVE-2022-45406 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongor5.0.0-rc5</b></p></summary> <p> <p>The MongoDB Database</p> <p>Library home page: <a href=https://github.com/mongodb/mongo.git>https://github.com/mongodb/mongo.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/mozjs-60/extract/js/src/vm/Realm.cpp</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/mozjs-60/extract/js/src/vm/Realm.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> If an out-of-memory condition occurred when creating a JavaScript global, a JavaScript realm may be deleted while references to it lived on in a BaseShape. This could lead to a use-after-free causing a potentially exploitable crash. This vulnerability affects Firefox ESR < 102.5, Thunderbird < 102.5, and Firefox < 107. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45406>CVE-2022-45406</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-45406 (Critical) detected in mongor5.0.0-rc5 - autoclosed - ## CVE-2022-45406 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongor5.0.0-rc5</b></p></summary> <p> <p>The MongoDB Database</p> <p>Library home page: <a href=https://github.com/mongodb/mongo.git>https://github.com/mongodb/mongo.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/mozjs-60/extract/js/src/vm/Realm.cpp</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/mozjs-60/extract/js/src/vm/Realm.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> If an out-of-memory condition occurred when creating a JavaScript global, a JavaScript realm may be deleted while references to it lived on in a BaseShape. This could lead to a use-after-free causing a potentially exploitable crash. This vulnerability affects Firefox ESR < 102.5, Thunderbird < 102.5, and Firefox < 107. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45406>CVE-2022-45406</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve critical detected in autoclosed cve critical severity vulnerability vulnerable library the mongodb database library home page a href found in base branch master vulnerable source files src third party mozjs extract js src vm realm cpp src third party mozjs extract js src vm realm cpp vulnerability details if an out of memory condition occurred when creating a javascript global a javascript realm may be deleted while references to it lived on in a baseshape this could lead to a use after free causing a potentially exploitable crash this vulnerability affects firefox esr thunderbird and firefox publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with mend
0
6,221
8,919,700,403
IssuesEvent
2019-01-21 02:17:53
Eason0731/MyPython
https://api.github.com/repos/Eason0731/MyPython
closed
Add game Tennis Titans backup and put back function on BackupArchiveFiles.py
requirement
Path of game Tennis Titans's archive file is: C:\Users\Administrator\AppData\Roaming\Macromedia\Director MX 2004\Tennis Titans sourceTennisTitans = os.path.join(os.environ['appdata'],'Macromedia','Director MX 2004','Tennis Titans') Just should add this function on BackupArchiveFiles.py before re-install Windows system. Then put back it.
1.0
Add game Tennis Titans backup and put back function on BackupArchiveFiles.py - Path of game Tennis Titans's archive file is: C:\Users\Administrator\AppData\Roaming\Macromedia\Director MX 2004\Tennis Titans sourceTennisTitans = os.path.join(os.environ['appdata'],'Macromedia','Director MX 2004','Tennis Titans') Just should add this function on BackupArchiveFiles.py before re-install Windows system. Then put back it.
non_infrastructure
add game tennis titans backup and put back function on backuparchivefiles py path of game tennis titans s archive file is c users administrator appdata roaming macromedia director mx tennis titans sourcetennistitans os path join os environ macromedia director mx tennis titans just should add this function on backuparchivefiles py before re install windows system then put back it
0
60,469
17,023,433,709
IssuesEvent
2021-07-03 02:00:42
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Track areas need to be rendered
Component: mapnik Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 10.29am, Thursday, 2nd July 2009]** I have done some mapping for the Royal Norfolk Show, visible at: http://www.openstreetmap.org/?lat=52.64898&lon=1.17735&zoom=16&layers=B000FTF This mainly consists of highway=track, tracktype=grade1; which renders fine. However, closed areas of the same surface tagged with area=yes are not rendered in Mapnik currently - though they appear in Osmarender.
1.0
Track areas need to be rendered - **[Submitted to the original trac issue database at 10.29am, Thursday, 2nd July 2009]** I have done some mapping for the Royal Norfolk Show, visible at: http://www.openstreetmap.org/?lat=52.64898&lon=1.17735&zoom=16&layers=B000FTF This mainly consists of highway=track, tracktype=grade1; which renders fine. However, closed areas of the same surface tagged with area=yes are not rendered in Mapnik currently - though they appear in Osmarender.
non_infrastructure
track areas need to be rendered i have done some mapping for the royal norfolk show visible at this mainly consists of highway track tracktype which renders fine however closed areas of the same surface tagged with area yes are not rendered in mapnik currently though they appear in osmarender
0
5,365
5,621,904,148
IssuesEvent
2017-04-04 11:19:30
phpmyadmin/phpmyadmin
https://api.github.com/repos/phpmyadmin/phpmyadmin
closed
Upload script uploads all -notes.html files
infrastructure
When calling `./scripts/upload-release 4.0.10.20 release/`, and with the 4.0.10.20 and 4.7.0 files both existing in `./release/`, this happens: phpMyAdmin/4.0.10.20/phpMyAdmin-4.0.10.20-source.tar.xz.sha256 phpMyAdmin/4.7.0-rc1/ phpMyAdmin/4.7.0-rc1/phpMyAdmin-4.7.0-rc1-notes.html
1.0
Upload script uploads all -notes.html files - When calling `./scripts/upload-release 4.0.10.20 release/`, and with the 4.0.10.20 and 4.7.0 files both existing in `./release/`, this happens: phpMyAdmin/4.0.10.20/phpMyAdmin-4.0.10.20-source.tar.xz.sha256 phpMyAdmin/4.7.0-rc1/ phpMyAdmin/4.7.0-rc1/phpMyAdmin-4.7.0-rc1-notes.html
infrastructure
upload script uploads all notes html files when calling scripts upload release release and with the and files both existing in release this happens phpmyadmin phpmyadmin source tar xz phpmyadmin phpmyadmin phpmyadmin notes html
1
28,157
23,062,709,504
IssuesEvent
2022-07-25 11:21:39
WordPress/performance
https://api.github.com/repos/WordPress/performance
opened
Link to new Persistent Object Caching documentation from Autoloaded Options health check
[Type] Documentation Infrastructure
As discussed in #234 and being reviewed in #448, once the new Persistent Object Caching section is ready and added, add a link to it in the [autoloaded options](https://wordpress.org/support/article/optimization/#autoloaded-options) section: Add a third one-sentence paragraph: "If you use a Persistent Object Cache, options (whether autoloaded or not) load faster and more efficiently." and link "Persistent Object Cache" to the new guide section.
1.0
Link to new Persistent Object Caching documentation from Autoloaded Options health check - As discussed in #234 and being reviewed in #448, once the new Persistent Object Caching section is ready and added, add a link to it in the [autoloaded options](https://wordpress.org/support/article/optimization/#autoloaded-options) section: Add a third one-sentence paragraph: "If you use a Persistent Object Cache, options (whether autoloaded or not) load faster and more efficiently." and link "Persistent Object Cache" to the new guide section.
infrastructure
link to new persistent object caching documentation from autoloaded options health check as discussed in and being reviewed in once the new persistent object caching section is ready and added add a link to it in the section add a third one sentence paragraph if you use a persistent object cache options whether autoloaded or not load faster and more efficiently and link persistent object cache to the new guide section
1
22,021
14,966,054,217
IssuesEvent
2021-01-27 14:09:02
aguirre-lab/ml4c3
https://api.github.com/repos/aguirre-lab/ml4c3
closed
Tensorization checks if desired file locations are mounted
infrastructure 🚇
## What and why To avoid a script running and not throwing an error when the user forgot to mount the required network share. ## Solution(s) Check if desired file locations are mounted. If not, throw error. ## Acceptance criteria If `tensorize` requires data from a specific mounted network share that is not mounted during runtime, a descriptive error is thrown. ## Blocked by or pending
1.0
Tensorization checks if desired file locations are mounted - ## What and why To avoid a script running and not throwing an error when the user forgot to mount the required network share. ## Solution(s) Check if desired file locations are mounted. If not, throw error. ## Acceptance criteria If `tensorize` requires data from a specific mounted network share that is not mounted during runtime, a descriptive error is thrown. ## Blocked by or pending
infrastructure
tensorization checks if desired file locations are mounted what and why to avoid a script running and not throwing an error when the user forgot to mount the required network share solution s check if desired file locations are mounted if not throw error acceptance criteria if tensorize requires data from a specific mounted network share that is not mounted during runtime a descriptive error is thrown blocked by or pending
1
510
2,743,336,562
IssuesEvent
2015-04-21 21:09:06
dinyar/uGMTfirmware
https://api.github.com/repos/dinyar/uGMTfirmware
closed
Modify build script to new build strategy
infrastructure
The build script from #183 has to be modified as the post implementation physical optimization step isn't necessary anymore. In the current form it fails the build in all cases.
1.0
Modify build script to new build strategy - The build script from #183 has to be modified as the post implementation physical optimization step isn't necessary anymore. In the current form it fails the build in all cases.
infrastructure
modify build script to new build strategy the build script from has to be modified as the post implementation physical optimization step isn t necessary anymore in the current form it fails the build in all cases
1
773,554
27,161,604,927
IssuesEvent
2023-02-17 12:20:27
tallyhowallet/extension
https://api.github.com/repos/tallyhowallet/extension
opened
Add NFTs tests - unit tests for NFTPreview
Type: Enhancement Status: Pending Priority: Medium
Test [NFTPreview](https://github.com/tallyhowallet/extension/blob/main/ui/components/NFTS_update/NFTPreview.tsx) component - displaying an image, network icon, floor price, owner - displaying placeholder image, unknown floor price - displaying description, contract and items in the collection - displaying cases with no description, no contract and unknown items in the collection - displaying too long description - displaying a list of attributes - displaying image for badge NFT
1.0
Add NFTs tests - unit tests for NFTPreview - Test [NFTPreview](https://github.com/tallyhowallet/extension/blob/main/ui/components/NFTS_update/NFTPreview.tsx) component - displaying an image, network icon, floor price, owner - displaying placeholder image, unknown floor price - displaying description, contract and items in the collection - displaying cases with no description, no contract and unknown items in the collection - displaying too long description - displaying a list of attributes - displaying image for badge NFT
non_infrastructure
add nfts tests unit tests for nftpreview test component displaying an image network icon floor price owner displaying placeholder image unknown floor price displaying description contract and items in the collection displaying cases with no description no contract and unknown items in the collection displaying too long description displaying a list of attributes displaying image for badge nft
0
569,398
17,013,559,478
IssuesEvent
2021-07-02 08:50:15
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
Allow login to the wiki using an OSM account
Component: wiki Priority: minor Type: enhancement
**[Submitted to the original trac issue database at 2.53pm, Saturday, 9th August 2008]** I'm new to the community and I came across a stumbling block last night. Basically I signed up for an account and then the next day I couldnt log in again. I tried resetting my password and all sorts and nothing worked I just couldnt get signed into the page. After an hour of getting really pissed off with the whole project I came to realise... you need two separate accounts, one for the map and one for the wiki, I had signed up for one but not the other. Is there no way you can have the same log in details be for both? I groaned at the prospect of having to create a 3rd account just to post here and suggest this, but it turns out the map password works here also, so in that case why can't it work for the wiki? One login to rule them all! In fact, how about supporting openID that way I dont even need to sign up for any accounts! Keep up the good work guys.
1.0
Allow login to the wiki using an OSM account - **[Submitted to the original trac issue database at 2.53pm, Saturday, 9th August 2008]** I'm new to the community and I came across a stumbling block last night. Basically I signed up for an account and then the next day I couldnt log in again. I tried resetting my password and all sorts and nothing worked I just couldnt get signed into the page. After an hour of getting really pissed off with the whole project I came to realise... you need two separate accounts, one for the map and one for the wiki, I had signed up for one but not the other. Is there no way you can have the same log in details be for both? I groaned at the prospect of having to create a 3rd account just to post here and suggest this, but it turns out the map password works here also, so in that case why can't it work for the wiki? One login to rule them all! In fact, how about supporting openID that way I dont even need to sign up for any accounts! Keep up the good work guys.
non_infrastructure
allow login to the wiki using an osm account i m new to the community and i came across a stumbling block last night basically i signed up for an account and then the next day i couldnt log in again i tried resetting my password and all sorts and nothing worked i just couldnt get signed into the page after an hour of getting really pissed off with the whole project i came to realise you need two separate accounts one for the map and one for the wiki i had signed up for one but not the other is there no way you can have the same log in details be for both i groaned at the prospect of having to create a account just to post here and suggest this but it turns out the map password works here also so in that case why can t it work for the wiki one login to rule them all in fact how about supporting openid that way i dont even need to sign up for any accounts keep up the good work guys
0
35,694
32,042,521,445
IssuesEvent
2023-09-22 20:44:12
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
opened
Provide more detail when a user tries to log in with a renamed account
[Feature] User & Account Settings (/me) [Type] Feature Request [Product] WordPress.com [Feature Group] Other Infrastructure
### What When a user tries to log in with the username for a renamed account, they currently just receive a message that their account was closed. Instead, for users whose account was renamed (as opposed to actually closed accounts), could we mention that the account was renamed? ### Why This is a low impact issue, but we do receive several emails a month from frustrated users who cannot understand why their account shows as closed when they try to log in using their old username. ### How _No response_
1.0
Provide more detail when a user tries to log in with a renamed account - ### What When a user tries to log in with the username for a renamed account, they currently just receive a message that their account was closed. Instead, for users whose account was renamed (as opposed to actually closed accounts), could we mention that the account was renamed? ### Why This is a low impact issue, but we do receive several emails a month from frustrated users who cannot understand why their account shows as closed when they try to log in using their old username. ### How _No response_
infrastructure
provide more detail when a user tries to log in with a renamed account what when a user tries to log in with the username for a renamed account they currently just receive a message that their account was closed instead for users whose account was renamed as opposed to actually closed accounts could we mention that the account was renamed why this is a low impact issue but we do receive several emails a month from frustrated users who cannot understand why their account shows as closed when they try to log in using their old username how no response
1
224,300
24,769,751,155
IssuesEvent
2022-10-23 01:22:37
turkdevops/grafana
https://api.github.com/repos/turkdevops/grafana
opened
CVE-2022-37598 (High) detected in uglify-js-3.4.10.tgz, uglify-js-3.6.2.tgz
security vulnerability
## CVE-2022-37598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>uglify-js-3.4.10.tgz</b>, <b>uglify-js-3.6.2.tgz</b></p></summary> <p> <details><summary><b>uglify-js-3.4.10.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.10.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.10.tgz</a></p> <p> Dependency Hierarchy: - @grafana/toolkit-6.7.0-pre.tgz (Root Library) - html-webpack-plugin-3.2.0.tgz - html-minifier-3.5.21.tgz - :x: **uglify-js-3.4.10.tgz** (Vulnerable Library) </details> <details><summary><b>uglify-js-3.6.2.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.2.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.2.tgz</a></p> <p> Dependency Hierarchy: - @grafana/toolkit-6.7.0-pre.tgz (Root Library) - jest-cli-24.9.0.tgz - core-24.9.0.tgz - reporters-24.9.0.tgz - istanbul-reports-2.2.6.tgz - handlebars-4.4.3.tgz - :x: **uglify-js-3.6.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p> <p>Found in base branch: <b>datasource-meta</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: uglify-js - 3.13.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-37598 (High) detected in uglify-js-3.4.10.tgz, uglify-js-3.6.2.tgz - ## CVE-2022-37598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>uglify-js-3.4.10.tgz</b>, <b>uglify-js-3.6.2.tgz</b></p></summary> <p> <details><summary><b>uglify-js-3.4.10.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.10.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.10.tgz</a></p> <p> Dependency Hierarchy: - @grafana/toolkit-6.7.0-pre.tgz (Root Library) - html-webpack-plugin-3.2.0.tgz - html-minifier-3.5.21.tgz - :x: **uglify-js-3.4.10.tgz** (Vulnerable Library) </details> <details><summary><b>uglify-js-3.6.2.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.2.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.2.tgz</a></p> <p> Dependency Hierarchy: - @grafana/toolkit-6.7.0-pre.tgz (Root Library) - jest-cli-24.9.0.tgz - core-24.9.0.tgz - reporters-24.9.0.tgz - istanbul-reports-2.2.6.tgz - handlebars-4.4.3.tgz - :x: **uglify-js-3.6.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p> <p>Found in base branch: <b>datasource-meta</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: uglify-js - 3.13.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in uglify js tgz uglify js tgz cve high severity vulnerability vulnerable libraries uglify js tgz uglify js tgz uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href dependency hierarchy grafana toolkit pre tgz root library html webpack plugin tgz html minifier tgz x uglify js tgz vulnerable library uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href dependency hierarchy grafana toolkit pre tgz root library jest cli tgz core tgz reporters tgz istanbul reports tgz handlebars tgz x uglify js tgz vulnerable library found in head commit a href found in base branch datasource meta vulnerability details prototype pollution vulnerability in function defnode in ast js in mishoo uglifyjs via the name variable in ast js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution uglify js step up your open source security game with mend
0
26,311
19,984,981,230
IssuesEvent
2022-01-30 14:19:58
planus-org/planus
https://api.github.com/repos/planus-org/planus
closed
[Rust] Enforce MSRV in CI
testing rust-backend infrastructure
If we have a minimum supported rust version of 1.57, we should also test in CI that we support that version.
1.0
[Rust] Enforce MSRV in CI - If we have a minimum supported rust version of 1.57, we should also test in CI that we support that version.
infrastructure
enforce msrv in ci if we have a minimum supported rust version of we should also test in ci that we support that version
1
15,333
11,458,841,927
IssuesEvent
2020-02-07 05:05:12
GoogleChrome/workbox
https://api.github.com/repos/GoogleChrome/workbox
opened
Linting for theTypeScript code
Infrastructure TypeScript
We never added in linting to the `.ts` files in our codebase. It looks like https://github.com/typescript-eslint/typescript-eslint is the standard package to use for that in 2020, though I'd assume we'll have to work on finding the best ruleset.
1.0
Linting for theTypeScript code - We never added in linting to the `.ts` files in our codebase. It looks like https://github.com/typescript-eslint/typescript-eslint is the standard package to use for that in 2020, though I'd assume we'll have to work on finding the best ruleset.
infrastructure
linting for thetypescript code we never added in linting to the ts files in our codebase it looks like is the standard package to use for that in though i d assume we ll have to work on finding the best ruleset
1
317,734
23,686,260,588
IssuesEvent
2022-08-29 06:38:31
eclipse-platform/eclipse.platform.swt
https://api.github.com/repos/eclipse-platform/eclipse.platform.swt
closed
SWT Javadoc bash for 4.25
documentation
**Describe the bug** Run Javadoc bash tool on SWT sources. **Expected behavior** Javadoc should be consistent in SWT code base for all platforms. **Environment:** 1. Select the platform(s) on which the behavior is seen: - - [x] All OS - - [ ] Windows - - [ ] Linux - - [ ] macOS 2. Additional OS info (e.g. OS version, Linux Desktop, etc) 3. JRE/JDK version **Version since** 4.25 **Workaround (or) Additional context** Add any other context about the problem here. Any known workarounds for the problem?
1.0
SWT Javadoc bash for 4.25 - **Describe the bug** Run Javadoc bash tool on SWT sources. **Expected behavior** Javadoc should be consistent in SWT code base for all platforms. **Environment:** 1. Select the platform(s) on which the behavior is seen: - - [x] All OS - - [ ] Windows - - [ ] Linux - - [ ] macOS 2. Additional OS info (e.g. OS version, Linux Desktop, etc) 3. JRE/JDK version **Version since** 4.25 **Workaround (or) Additional context** Add any other context about the problem here. Any known workarounds for the problem?
non_infrastructure
swt javadoc bash for describe the bug run javadoc bash tool on swt sources expected behavior javadoc should be consistent in swt code base for all platforms environment select the platform s on which the behavior is seen all os windows linux macos additional os info e g os version linux desktop etc jre jdk version version since workaround or additional context add any other context about the problem here any known workarounds for the problem
0
20,210
13,756,969,446
IssuesEvent
2020-10-06 20:48:57
nwfsc-fram/boatnet
https://api.github.com/repos/nwfsc-fram/boatnet
closed
Create Jenkins Boatnet-Auth Health Check
Prj:infrastructure
The auth server appeared to stop working on 8/13/2019 for some unknown reason. Please implement a jenkins boatnet-auth health check once an hour to ensure that the system is operational.
1.0
Create Jenkins Boatnet-Auth Health Check - The auth server appeared to stop working on 8/13/2019 for some unknown reason. Please implement a jenkins boatnet-auth health check once an hour to ensure that the system is operational.
infrastructure
create jenkins boatnet auth health check the auth server appeared to stop working on for some unknown reason please implement a jenkins boatnet auth health check once an hour to ensure that the system is operational
1
21,050
14,289,991,220
IssuesEvent
2020-11-23 20:09:21
Eastrall/Rhisis
https://api.github.com/repos/Eastrall/Rhisis
opened
Replace CoreServer with a messaging queue system
feature-request infrastructure refactoring srv: core
In the actual solution, the servers communicate using a custom socket server on `LoginServer` and `ClusterServer`. These servers are used to broadcast messages to the connected `WorldServers`. Instead of maintaining a custom solution, I believe we should use an existing message queue that (in my opinion) suits perfectly what we want to achieve here. I propose to use RabbitMQ which is an open-source message queue system. When PR #487 will be merged, I will create a new docker container with this message queue server and provide an easy API to publish and subscribe to messages instead of using RabbitMQ native methods. This allows us to change the messaging queue system whenever we want without breaking any functionnal code.
1.0
Replace CoreServer with a messaging queue system - In the actual solution, the servers communicate using a custom socket server on `LoginServer` and `ClusterServer`. These servers are used to broadcast messages to the connected `WorldServers`. Instead of maintaining a custom solution, I believe we should use an existing message queue that (in my opinion) suits perfectly what we want to achieve here. I propose to use RabbitMQ which is an open-source message queue system. When PR #487 will be merged, I will create a new docker container with this message queue server and provide an easy API to publish and subscribe to messages instead of using RabbitMQ native methods. This allows us to change the messaging queue system whenever we want without breaking any functionnal code.
infrastructure
replace coreserver with a messaging queue system in the actual solution the servers communicate using a custom socket server on loginserver and clusterserver these servers are used to broadcast messages to the connected worldservers instead of maintaining a custom solution i believe we should use an existing message queue that in my opinion suits perfectly what we want to achieve here i propose to use rabbitmq which is an open source message queue system when pr will be merged i will create a new docker container with this message queue server and provide an easy api to publish and subscribe to messages instead of using rabbitmq native methods this allows us to change the messaging queue system whenever we want without breaking any functionnal code
1
13,157
10,131,821,434
IssuesEvent
2019-08-01 20:33:45
HumanCellAtlas/secondary-analysis
https://api.github.com/repos/HumanCellAtlas/secondary-analysis
reopened
Does the Matrix service have what it needs to serve mouse matrices
infrastructure
Create a test analysis bundle in integration env for Matrix Service testing. For Mouse Smartseq2. Share this bundle with Marcus Kinsella at CZI via email or HCA slack to see if the matrix service needs anything else to be able to run mouse data. If they don’t need anything, ask for feedback on mouse analysis outputs for Optimus and SS2. **Why?** The Matrix service may need more information, or changes to the mouse data outputs **Where to start?** Talk to Nick/Kishori, and reach out to Marcus via Slack. Check if there is SS2 mouse example output bundles in a DCP environment. Work with Lantern to create them if there is not, it can be done before validation is complete, as long as it’s known it cannot be merged into Prod until it is validated]. **When is it done?** When the matrix service can support mouse data outputs, or when we have a list of task needed to complete matrix service support. ┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-357)
1.0
Does the Matrix service have what it needs to serve mouse matrices - Create a test analysis bundle in integration env for Matrix Service testing. For Mouse Smartseq2. Share this bundle with Marcus Kinsella at CZI via email or HCA slack to see if the matrix service needs anything else to be able to run mouse data. If they don’t need anything, ask for feedback on mouse analysis outputs for Optimus and SS2. **Why?** The Matrix service may need more information, or changes to the mouse data outputs **Where to start?** Talk to Nick/Kishori, and reach out to Marcus via Slack. Check if there is SS2 mouse example output bundles in a DCP environment. Work with Lantern to create them if there is not, it can be done before validation is complete, as long as it’s known it cannot be merged into Prod until it is validated]. **When is it done?** When the matrix service can support mouse data outputs, or when we have a list of task needed to complete matrix service support. ┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-357)
infrastructure
does the matrix service have what it needs to serve mouse matrices create a test analysis bundle in integration env for matrix service testing for mouse share this bundle with marcus kinsella at czi via email or hca slack to see if the matrix service needs anything else to be able to run mouse data if they don’t need anything ask for feedback on mouse analysis outputs for optimus and why the matrix service may need more information or changes to the mouse data outputs where to start talk to nick kishori and reach out to marcus via slack check if there is mouse example output bundles in a dcp environment work with lantern to create them if there is not it can be done before validation is complete as long as it’s known it cannot be merged into prod until it is validated when is it done when the matrix service can support mouse data outputs or when we have a list of task needed to complete matrix service support ┆issue is synchronized with this
1
6,670
6,551,080,040
IssuesEvent
2017-09-05 13:37:41
dzhw/metadatamanagement
https://api.github.com/repos/dzhw/metadatamanagement
closed
Connect variables that have been derived from each other
category:variablemanagement points:13 prio:1 scope:project infrastructure type:backlog item
I think we talked about this several times but I don't remember why we did not do this, so sorry in advance: I think users will appreciate it if variables that only differ by suffix were conntected, e.g.: Someone looks at the variable questionnaire and is interested in "1. Qualifikation: Hochschulart" which is bstu09_a, thus not accessible. As users do not (yet) know how the variable suffix works it could be very helpful to display other variables that come from bstu09_a (and are accessible) which are: - bstu09_g1r -> 1. Qualifikation: Bundesland der Hochschule - bstu09_g2a -> 1. Qualifikation: Staat einer ausl. Hochschule - bstu09_g3r -> 1. Qualifikation: Staat einer ausl. Hochschule - bstu09_g4o -> 1. Qualifikation: Hochschulart - bstu09_g5r -> 1. Qualifikation: Hochschulart (FH/Uni; ohne Saarland) Could these be "cross-listed" in all these variables in "related variables"?
1.0
Connect variables that have been derived from each other - I think we talked about this several times but I don't remember why we did not do this, so sorry in advance: I think users will appreciate it if variables that only differ by suffix were conntected, e.g.: Someone looks at the variable questionnaire and is interested in "1. Qualifikation: Hochschulart" which is bstu09_a, thus not accessible. As users do not (yet) know how the variable suffix works it could be very helpful to display other variables that come from bstu09_a (and are accessible) which are: - bstu09_g1r -> 1. Qualifikation: Bundesland der Hochschule - bstu09_g2a -> 1. Qualifikation: Staat einer ausl. Hochschule - bstu09_g3r -> 1. Qualifikation: Staat einer ausl. Hochschule - bstu09_g4o -> 1. Qualifikation: Hochschulart - bstu09_g5r -> 1. Qualifikation: Hochschulart (FH/Uni; ohne Saarland) Could these be "cross-listed" in all these variables in "related variables"?
infrastructure
connect variables that have been derived from each other i think we talked about this several times but i don t remember why we did not do this so sorry in advance i think users will appreciate it if variables that only differ by suffix were conntected e g someone looks at the variable questionnaire and is interested in qualifikation hochschulart which is a thus not accessible as users do not yet know how the variable suffix works it could be very helpful to display other variables that come from a and are accessible which are qualifikation bundesland der hochschule qualifikation staat einer ausl hochschule qualifikation staat einer ausl hochschule qualifikation hochschulart qualifikation hochschulart fh uni ohne saarland could these be cross listed in all these variables in related variables
1
169,949
13,166,764,149
IssuesEvent
2020-08-11 09:06:33
WoWManiaUK/Blackwing-Lair
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
closed
[Quest] [Healing in a flash]
Confirmed By Tester Fixed Confirmed Fixed in Dev
**Links:** http://cata.cavernoftime.com/quest=27066 from WoWHead or our Armory **What is happening:** when you try using Flash Heal on the wounded braves their health goes up but you don't receive credit for it. **What should happen:** **Is it crashing the server?:** **Other Information:**
1.0
[Quest] [Healing in a flash] - **Links:** http://cata.cavernoftime.com/quest=27066 from WoWHead or our Armory **What is happening:** when you try using Flash Heal on the wounded braves their health goes up but you don't receive credit for it. **What should happen:** **Is it crashing the server?:** **Other Information:**
non_infrastructure
links from wowhead or our armory what is happening when you try using flash heal on the wounded braves their health goes up but you don t receive credit for it what should happen is it crashing the server other information
0
25,909
19,400,277,133
IssuesEvent
2021-12-19 03:08:32
legal-network/scotus-analysis
https://api.github.com/repos/legal-network/scotus-analysis
closed
Convert network initialization queries to SQLAlchemy
infrastructure
Convert the queries in `network_edge_list.py` and `citation_network.py` and do surrounding work that makes that feasible
1.0
Convert network initialization queries to SQLAlchemy - Convert the queries in `network_edge_list.py` and `citation_network.py` and do surrounding work that makes that feasible
infrastructure
convert network initialization queries to sqlalchemy convert the queries in network edge list py and citation network py and do surrounding work that makes that feasible
1
15,447
11,516,500,281
IssuesEvent
2020-02-14 05:19:56
servo/servo
https://api.github.com/repos/servo/servo
closed
Docker image creation is busted
A-infrastructure I-bustage
The docker image was created as part of the WPT update job today, and it failed with: ``` Collecting virtualenv Downloading https://files.pythonhosted.org/packages/18/cb/9799b00a192391cab56f1e6400bfdeb331afff7527e18dc3df25543782ff/virtualenv-20.0.3-py2.py3-none-any.whl (4.6MB) Collecting filelock<4,>=3.0.0 (from virtualenv) Downloading https://files.pythonhosted.org/packages/14/ec/6ee2168387ce0154632f856d5cc5592328e9cf93127c5c9aeca92c8c16cb/filelock-3.0.12.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> ImportError: No module named setuptools ``` Virtualenv v20 was release in the past month, which is the last time the image was recreated.
1.0
Docker image creation is busted - The docker image was created as part of the WPT update job today, and it failed with: ``` Collecting virtualenv Downloading https://files.pythonhosted.org/packages/18/cb/9799b00a192391cab56f1e6400bfdeb331afff7527e18dc3df25543782ff/virtualenv-20.0.3-py2.py3-none-any.whl (4.6MB) Collecting filelock<4,>=3.0.0 (from virtualenv) Downloading https://files.pythonhosted.org/packages/14/ec/6ee2168387ce0154632f856d5cc5592328e9cf93127c5c9aeca92c8c16cb/filelock-3.0.12.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> ImportError: No module named setuptools ``` Virtualenv v20 was release in the past month, which is the last time the image was recreated.
infrastructure
docker image creation is busted the docker image was created as part of the wpt update job today and it failed with collecting virtualenv downloading collecting filelock from virtualenv downloading complete output from command python setup py egg info traceback most recent call last file line in importerror no module named setuptools virtualenv was release in the past month which is the last time the image was recreated
1
29,181
23,779,813,529
IssuesEvent
2022-09-02 02:37:19
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Can soil layers be deleted in the GUI?
bug interface/infrastructure
This is perhaps a silly question, but once soil layers have been added in the GUI, is there any way to delete them? Some of my attempts to do so resulted in exceptions and application crash.
1.0
Can soil layers be deleted in the GUI? - This is perhaps a silly question, but once soil layers have been added in the GUI, is there any way to delete them? Some of my attempts to do so resulted in exceptions and application crash.
infrastructure
can soil layers be deleted in the gui this is perhaps a silly question but once soil layers have been added in the gui is there any way to delete them some of my attempts to do so resulted in exceptions and application crash
1
88,411
8,140,788,092
IssuesEvent
2018-08-20 22:28:56
InfiniteAdventures/ia-trilogy
https://api.github.com/repos/InfiniteAdventures/ia-trilogy
closed
Quelltext-Fehler: Dort steht "\^C". Gemeint ist: "\^{}C"
Sonstiges Problem Testdruck-Fehler
Kaum zu sehen, aber natürlich soll da kein Zirkumflex über dem C erscheinen. Gemeint ist die übliche Kommandozeilen-Schreibweise für "Strg+C", also "^C".
1.0
Quelltext-Fehler: Dort steht "\^C". Gemeint ist: "\^{}C" - Kaum zu sehen, aber natürlich soll da kein Zirkumflex über dem C erscheinen. Gemeint ist die übliche Kommandozeilen-Schreibweise für "Strg+C", also "^C".
non_infrastructure
quelltext fehler dort steht c gemeint ist c kaum zu sehen aber natürlich soll da kein zirkumflex über dem c erscheinen gemeint ist die übliche kommandozeilen schreibweise für strg c also c
0
44,652
12,308,034,027
IssuesEvent
2020-05-12 06:19:45
jankeirse/tvhgooglemapi
https://api.github.com/repos/jankeirse/tvhgooglemapi
closed
BCC addresses are applied to the To field for recipiets
Priority-Medium Type-Defect auto-migrated wontfix
``` What steps will reproduce the problem? 1. Access DB using the DoCmd.SendObject 2. Have list of emails in the BCC paramenter 3. http://msdn.microsoft.com/en-us/library/office/ff197046%28v=office.15%29.aspx What is the expected output? What do you see instead? Since the addresses are listed as BCC, I would expected them to be added to the BCC field. This works when I use Lotus Notes. What version of the product are you using? On what operating system? 0.8.2 on Win 7 Please provide any additional information below. Jan, it's Ben in olathe. Give a ring if you have ant other questions. ``` Original issue reported on code.google.com by `ben.flem...@tvh.com` on 12 Dec 2014 at 8:33
1.0
BCC addresses are applied to the To field for recipiets - ``` What steps will reproduce the problem? 1. Access DB using the DoCmd.SendObject 2. Have list of emails in the BCC paramenter 3. http://msdn.microsoft.com/en-us/library/office/ff197046%28v=office.15%29.aspx What is the expected output? What do you see instead? Since the addresses are listed as BCC, I would expected them to be added to the BCC field. This works when I use Lotus Notes. What version of the product are you using? On what operating system? 0.8.2 on Win 7 Please provide any additional information below. Jan, it's Ben in olathe. Give a ring if you have ant other questions. ``` Original issue reported on code.google.com by `ben.flem...@tvh.com` on 12 Dec 2014 at 8:33
non_infrastructure
bcc addresses are applied to the to field for recipiets what steps will reproduce the problem access db using the docmd sendobject have list of emails in the bcc paramenter what is the expected output what do you see instead since the addresses are listed as bcc i would expected them to be added to the bcc field this works when i use lotus notes what version of the product are you using on what operating system on win please provide any additional information below jan it s ben in olathe give a ring if you have ant other questions original issue reported on code google com by ben flem tvh com on dec at
0
16,166
4,021,181,253
IssuesEvent
2016-05-16 21:01:47
saltstack/salt
https://api.github.com/repos/saltstack/salt
closed
Changes dictionary return should be mentioned in test state docs
Bug Documentation Fixed Pending Verification High Severity P3 TEAM Core
The [Test State](https://docs.saltstack.com/en/latest/ref/states/writing.html#test-state) section of the [Writing States](https://docs.saltstack.com/en/latest/ref/states/writing.html) doc doesn't contain an example of the changes dictionary. The changes dictionary often contains a lot of helpful information that should be included in the `test=true` functionality when people are writing states. The example should contain this information.
1.0
Changes dictionary return should be mentioned in test state docs - The [Test State](https://docs.saltstack.com/en/latest/ref/states/writing.html#test-state) section of the [Writing States](https://docs.saltstack.com/en/latest/ref/states/writing.html) doc doesn't contain an example of the changes dictionary. The changes dictionary often contains a lot of helpful information that should be included in the `test=true` functionality when people are writing states. The example should contain this information.
non_infrastructure
changes dictionary return should be mentioned in test state docs the section of the doc doesn t contain an example of the changes dictionary the changes dictionary often contains a lot of helpful information that should be included in the test true functionality when people are writing states the example should contain this information
0