Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
13,979 | 10,569,163,098 | IssuesEvent | 2019-10-06 17:46:44 | visualworks/portaljal.com.br | https://api.github.com/repos/visualworks/portaljal.com.br | opened | Retry failed network request | bug infrastructure thirdparty ux | When a network request fails either to GlobalBus or to the user's base, there's no option to retrigger unless doing a browser refresh, therefore retriggering all the request. We want to retrigger only the failed refresh. | 1.0 | Retry failed network request - When a network request fails either to GlobalBus or to the user's base, there's no option to retrigger unless doing a browser refresh, therefore retriggering all the request. We want to retrigger only the failed refresh. | infrastructure | retry failed network request when a network request fails either to globalbus or to the user s base there s no option to retrigger unless doing a browser refresh therefore retriggering all the request we want to retrigger only the failed refresh | 1 |
11,269 | 9,040,170,751 | IssuesEvent | 2019-02-10 14:09:29 | coq/coq | https://api.github.com/repos/coq/coq | opened | dune and coqtop.byte | kind: bug kind: infrastructure | <!-- Thank you for reporting a bug to Coq! -->
#### Description of the problem
<!-- If you can, it's helpful to provide self-contained example of some code
that reproduces the bug. If not, a link to a larger example is also helpful. -->
I was trying to use coqtop.byte built with dune to debug something in equations and failed.
First issue is that just running coqtop.byte failed, one has to
```
export LD_LIBRARY_PATH=PATH_TO_COQ/_build/install/default//lib/stublibs:$LD_LIBRARY_PATH
```
For this to run, otherwise the dynamic library linker fails to find a DLL.
After this, processing works, and I can drop to the toplevel, however `#use "include";;` fails, and after adding the PATH_TO_COQ/dev directory to the source files it succeeds but cannot find the object files it needs. Adding #directory "PATH_TO_COQ/_build/default/dev" it still fails with:
```
File "/Users/sozeau/dev/coq/coq/_build/default/dev/top_printers.ml", line 13, characters 5-10:
Error: Unbound module Sorts
Hint: Did you mean Sort?
File "/Users/sozeau/dev/coq/coq/_build/default/dev/vm_printers.ml", line 2, characters 5-9:
Error: Unbound module Term
Unbound value ppid.
...
```
| 1.0 | dune and coqtop.byte - <!-- Thank you for reporting a bug to Coq! -->
#### Description of the problem
<!-- If you can, it's helpful to provide self-contained example of some code
that reproduces the bug. If not, a link to a larger example is also helpful. -->
I was trying to use coqtop.byte built with dune to debug something in equations and failed.
First issue is that just running coqtop.byte failed, one has to
```
export LD_LIBRARY_PATH=PATH_TO_COQ/_build/install/default//lib/stublibs:$LD_LIBRARY_PATH
```
For this to run, otherwise the dynamic library linker fails to find a DLL.
After this, processing works, and I can drop to the toplevel, however `#use "include";;` fails, and after adding the PATH_TO_COQ/dev directory to the source files it succeeds but cannot find the object files it needs. Adding #directory "PATH_TO_COQ/_build/default/dev" it still fails with:
```
File "/Users/sozeau/dev/coq/coq/_build/default/dev/top_printers.ml", line 13, characters 5-10:
Error: Unbound module Sorts
Hint: Did you mean Sort?
File "/Users/sozeau/dev/coq/coq/_build/default/dev/vm_printers.ml", line 2, characters 5-9:
Error: Unbound module Term
Unbound value ppid.
...
```
| infrastructure | dune and coqtop byte description of the problem if you can it s helpful to provide self contained example of some code that reproduces the bug if not a link to a larger example is also helpful i was trying to use coqtop byte built with dune to debug something in equations and failed first issue is that just running coqtop byte failed one has to export ld library path path to coq build install default lib stublibs ld library path for this to run otherwise the dynamic library linker fails to find a dll after this processing works and i can drop to the toplevel however use include fails and after adding the path to coq dev directory to the source files it succeeds but cannot find the object files it needs adding directory path to coq build default dev it still fails with file users sozeau dev coq coq build default dev top printers ml line characters error unbound module sorts hint did you mean sort file users sozeau dev coq coq build default dev vm printers ml line characters error unbound module term unbound value ppid | 1 |
92,574 | 26,719,440,960 | IssuesEvent | 2023-01-29 00:03:11 | envoyproxy/envoy | https://api.github.com/repos/envoyproxy/envoy | closed | build/mobile: RBE fails with arm64 versions of Bazel on macOS | area/build stale mobile | *Title*: build/mobile: RBE fails with arm64 versions of Bazel on macOS
*Description*:
When building something in the `mobile/` workspace with an arm64 bazel binary on an M1 Mac with RBE enabled, the `protoc` with the wrong architecture is attempted to be run, leading to [this error](https://envoy.cluster.engflow.com/invocation/30cc6fcc-8c30-46b0-b616-48af66cec8a9#console):
> sandbox-exec: execvp() of 'bazel-out/.../com_google_protobuf/protoc' failed: Bad CPU type in executable
*Repro steps*:
On an M1 Mac with EngFlow certs installed at `/path/to/certs/engflow.crt` and `/path/to/certs/engflow.key`:
```
git clone https://github.com/envoyproxy/envoy.git
cd envoy/mobile
./bazelw build \
--tls_client_certificate=/path/to/certs/engflow.crt \
--tls_client_key=/path/to/certs/engflow.key \
--config remote-ci-macos \
//library/common:envoy_main_interface_lib
```
The build succeeds if this line gets hardcoded to `amd64`: https://github.com/envoyproxy/envoy/blob/367763e5c63c003e495981e98fd9967bbcb86d20/mobile/bazelw#L16
*Admin and Stats Output*:
N/A
*Config*:
N/A
*Logs*:
N/A
*Call Stack*:
N/A | 1.0 | build/mobile: RBE fails with arm64 versions of Bazel on macOS - *Title*: build/mobile: RBE fails with arm64 versions of Bazel on macOS
*Description*:
When building something in the `mobile/` workspace with an arm64 bazel binary on an M1 Mac with RBE enabled, the `protoc` with the wrong architecture is attempted to be run, leading to [this error](https://envoy.cluster.engflow.com/invocation/30cc6fcc-8c30-46b0-b616-48af66cec8a9#console):
> sandbox-exec: execvp() of 'bazel-out/.../com_google_protobuf/protoc' failed: Bad CPU type in executable
*Repro steps*:
On an M1 Mac with EngFlow certs installed at `/path/to/certs/engflow.crt` and `/path/to/certs/engflow.key`:
```
git clone https://github.com/envoyproxy/envoy.git
cd envoy/mobile
./bazelw build \
--tls_client_certificate=/path/to/certs/engflow.crt \
--tls_client_key=/path/to/certs/engflow.key \
--config remote-ci-macos \
//library/common:envoy_main_interface_lib
```
The build succeeds if this line gets hardcoded to `amd64`: https://github.com/envoyproxy/envoy/blob/367763e5c63c003e495981e98fd9967bbcb86d20/mobile/bazelw#L16
*Admin and Stats Output*:
N/A
*Config*:
N/A
*Logs*:
N/A
*Call Stack*:
N/A | non_infrastructure | build mobile rbe fails with versions of bazel on macos title build mobile rbe fails with versions of bazel on macos description when building something in the mobile workspace with an bazel binary on an mac with rbe enabled the protoc with the wrong architecture is attempted to be run leading to sandbox exec execvp of bazel out com google protobuf protoc failed bad cpu type in executable repro steps on an mac with engflow certs installed at path to certs engflow crt and path to certs engflow key git clone cd envoy mobile bazelw build tls client certificate path to certs engflow crt tls client key path to certs engflow key config remote ci macos library common envoy main interface lib the build succeeds if this line gets hardcoded to admin and stats output n a config n a logs n a call stack n a | 0 |
20,924 | 14,239,812,627 | IssuesEvent | 2020-11-18 20:44:07 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Cannot browse source code in Visual Studio | area-infrastructure | ### Describe the bug
When opening the source code from this repository in Visual Studio, there are two issues. First, the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red. Second, I repeatedly get modal Microsoft login dialogs that want me to login somewhere. When I do log in, it has no effect at all. Visual Studio is completely unusable with these dialogs popping up all the time!
### To Reproduce
1. Check out the source code from this repository.
2. Switch to the tag v3.1.6
3. Change global.json to match the installed .NET SDK version
4. Open src\Mvc\Mvc.sln in Visual Studio 2019 Community Edition 16.6.4
### Exceptions (if any)
This is the package manager output:
```
Alle Pakete sind bereits installiert. Nichts muss wiederhergestellt werden.
Verstrichene Zeit: 00:01:46.2440102
========== Abgeschlossen ==========
Fehler beim Wiederherstellen von NuGet-Paketen: Der Dienstindex für die Quelle "https://pkgs.dev.azure.com/dnceng/_packaging/darc-int-dotnet-extensions-be18161f/nuget/v3/index.json" konnte nicht geladen werden.
```
### Further technical details
dotnet --info:
```
.NET Core SDK (gemäß "global.json"):
Version: 3.1.302
Commit: 41faccf259
Laufzeitumgebung:
OS Name: Windows
OS Version: 10.0.19041
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\3.1.302\
Host (useful for support):
Version: 3.1.6
Commit: 3acd9b0cd1
.NET Core SDKs installed:
3.1.302 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
| 1.0 | Cannot browse source code in Visual Studio - ### Describe the bug
When opening the source code from this repository in Visual Studio, there are two issues. First, the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red. Second, I repeatedly get modal Microsoft login dialogs that want me to login somewhere. When I do log in, it has no effect at all. Visual Studio is completely unusable with these dialogs popping up all the time!
### To Reproduce
1. Check out the source code from this repository.
2. Switch to the tag v3.1.6
3. Change global.json to match the installed .NET SDK version
4. Open src\Mvc\Mvc.sln in Visual Studio 2019 Community Edition 16.6.4
### Exceptions (if any)
This is the package manager output:
```
Alle Pakete sind bereits installiert. Nichts muss wiederhergestellt werden.
Verstrichene Zeit: 00:01:46.2440102
========== Abgeschlossen ==========
Fehler beim Wiederherstellen von NuGet-Paketen: Der Dienstindex für die Quelle "https://pkgs.dev.azure.com/dnceng/_packaging/darc-int-dotnet-extensions-be18161f/nuget/v3/index.json" konnte nicht geladen werden.
```
### Further technical details
dotnet --info:
```
.NET Core SDK (gemäß "global.json"):
Version: 3.1.302
Commit: 41faccf259
Laufzeitumgebung:
OS Name: Windows
OS Version: 10.0.19041
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\3.1.302\
Host (useful for support):
Version: 3.1.6
Commit: 3acd9b0cd1
.NET Core SDKs installed:
3.1.302 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
| infrastructure | cannot browse source code in visual studio describe the bug when opening the source code from this repository in visual studio there are two issues first the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red second i repeatedly get modal microsoft login dialogs that want me to login somewhere when i do log in it has no effect at all visual studio is completely unusable with these dialogs popping up all the time to reproduce check out the source code from this repository switch to the tag change global json to match the installed net sdk version open src mvc mvc sln in visual studio community edition exceptions if any this is the package manager output alle pakete sind bereits installiert nichts muss wiederhergestellt werden verstrichene zeit abgeschlossen fehler beim wiederherstellen von nuget paketen der dienstindex für die quelle konnte nicht geladen werden further technical details dotnet info net core sdk gemäß global json version commit laufzeitumgebung os name windows os version os platform windows rid base path c program files dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app to install additional net core runtimes or sdks | 1 |
786,055 | 27,632,852,562 | IssuesEvent | 2023-03-10 12:16:26 | TESTARtool/TESTAR_dev | https://api.github.com/repos/TESTARtool/TESTAR_dev | closed | Update OrientDB CI Actions URL | type: technical feature request priority: high status: not started | Seems that OrientDB migrated the hosting of its databases from:
https://s3.us-east-2.amazonaws.com/orientdb3/releases/3.0.34
to:
https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.0.34/
Change the related Gradle Tasks to fix the CI Actions | 1.0 | Update OrientDB CI Actions URL - Seems that OrientDB migrated the hosting of its databases from:
https://s3.us-east-2.amazonaws.com/orientdb3/releases/3.0.34
to:
https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.0.34/
Change the related Gradle Tasks to fix the CI Actions | non_infrastructure | update orientdb ci actions url seems that orientdb migrated the hosting of its databases from to change the related gradle tasks to fix the ci actions | 0 |
297,071 | 25,597,536,646 | IssuesEvent | 2022-12-01 17:17:43 | elastic/kibana | https://api.github.com/repos/elastic/kibana | reopened | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js - logstash pipeline create new cancel button discards the pipeline and redirects to the list | blocker Feature:Logstash Pipelines Team:Logstash failed-test skipped-test v7.11.0 | A test failed on a tracked branch
```
Error: retry.try timeout: Error: expected [ { selected: false,
id: 'cuhvoemuemat',
description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.',
lastModified: 'A few seconds ago',
username: 'test_user' },
{ selected: false,
id: 'empty_pipeline_1',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_10',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_11',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_12',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_13',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_14',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_15',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_16',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_17',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_18',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_19',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_2',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_20',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_21',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_3',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_4',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_5',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_6',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_7',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' } ] to sort of equal [ { selected: false,
id: 'cuhvoemuemat',
description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.',
lastModified: 'A minute ago',
username: 'test_user' },
{ selected: false,
id: 'empty_pipeline_1',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_10',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_11',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_12',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_13',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_14',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_15',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_16',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_17',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_18',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_19',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_2',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_20',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_21',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_3',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_4',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_5',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_6',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_7',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' } ]
at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:244:8)
at retry.try (test/functional/apps/logstash/pipeline_create.js:98:35)
at process._tickCallback (internal/process/next_tick.js:68:7)
at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:68:13)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9586/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js","test.name":"logstash pipeline create new cancel button discards the pipeline and redirects to the list","test.failCount":3}} --> | 2.0 | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js - logstash pipeline create new cancel button discards the pipeline and redirects to the list - A test failed on a tracked branch
```
Error: retry.try timeout: Error: expected [ { selected: false,
id: 'cuhvoemuemat',
description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.',
lastModified: 'A few seconds ago',
username: 'test_user' },
{ selected: false,
id: 'empty_pipeline_1',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_10',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_11',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_12',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_13',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_14',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_15',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_16',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_17',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_18',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_19',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_2',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_20',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_21',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_3',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_4',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_5',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_6',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_7',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' } ] to sort of equal [ { selected: false,
id: 'cuhvoemuemat',
description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.',
lastModified: 'A minute ago',
username: 'test_user' },
{ selected: false,
id: 'empty_pipeline_1',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_10',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_11',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_12',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_13',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_14',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_15',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_16',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_17',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_18',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_19',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_2',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_20',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_21',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_3',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_4',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_5',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_6',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' },
{ selected: false,
id: 'empty_pipeline_7',
description: 'an empty pipeline',
lastModified: '3 years ago',
username: 'elastic' } ]
at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:244:8)
at retry.try (test/functional/apps/logstash/pipeline_create.js:98:35)
at process._tickCallback (internal/process/next_tick.js:68:7)
at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:68:13)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9586/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js","test.name":"logstash pipeline create new cancel button discards the pipeline and redirects to the list","test.failCount":3}} --> | non_infrastructure | failing test chrome x pack ui functional tests x pack test functional apps logstash pipeline create·js logstash pipeline create new cancel button discards the pipeline and redirects to the list a test failed on a tracked branch error retry try timeout error expected selected false id cuhvoemuemat description ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse lastmodified a few seconds ago username test user selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic to sort of equal selected false id cuhvoemuemat description ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse lastmodified a minute ago username test user selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic at assertion assert dev shm workspace parallel kibana packages kbn expect expect js at assertion eql dev shm workspace parallel kibana packages kbn expect expect js at retry try test functional apps logstash pipeline create js at process tickcallback internal process next tick js at onfailure dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts first failure | 0 |
9,087 | 7,807,788,640 | IssuesEvent | 2018-06-11 18:01:50 | ampproject/docs | https://api.github.com/repos/ampproject/docs | closed | Build fails on "Error 502" | P0: Critical Type: Bug Type: Site Infrastructure | This error started today and I'm not sure what the problem is. The 3 other parallel jobs worked.
```
Warning: Don't know how to categorize amp-subscriptions-google.md, skipping..
/home/travis/build/ampproject/docs/scripts/import_docs.js:32
throw err;
^
Error: Error 502
at ChildProcess.exithandler (child_process.js:275:12)
at emitTwo (events.js:126:13)
at ChildProcess.emit (events.js:214:7)
at maybeClose (internal/child_process.js:925:16)
at Socket.stream.socket.on (internal/child_process.js:346:11)
at emitOne (events.js:116:13)
at Socket.emit (events.js:211:7)
at Pipe._handle.close [as _onclose] (net.js:557:12)
[17:14:48] 'build' errored after 9.51 s
[17:14:48] The following tasks did not complete: optimize-images, sass
[17:14:48] Did you forget to signal async completion?
```
Log: https://travis-ci.org/ampproject/docs/jobs/390843966
I've restarted the job several times and it either just fails on the script or produces a 502 error with warnings about categorization --- which shouldn't occur.
I'm thinking it's flakiness but not sure.
Can't reproduce locally. | 1.0 | Build fails on "Error 502" - This error started today and I'm not sure what the problem is. The 3 other parallel jobs worked.
```
Warning: Don't know how to categorize amp-subscriptions-google.md, skipping..
/home/travis/build/ampproject/docs/scripts/import_docs.js:32
throw err;
^
Error: Error 502
at ChildProcess.exithandler (child_process.js:275:12)
at emitTwo (events.js:126:13)
at ChildProcess.emit (events.js:214:7)
at maybeClose (internal/child_process.js:925:16)
at Socket.stream.socket.on (internal/child_process.js:346:11)
at emitOne (events.js:116:13)
at Socket.emit (events.js:211:7)
at Pipe._handle.close [as _onclose] (net.js:557:12)
[17:14:48] 'build' errored after 9.51 s
[17:14:48] The following tasks did not complete: optimize-images, sass
[17:14:48] Did you forget to signal async completion?
```
Log: https://travis-ci.org/ampproject/docs/jobs/390843966
I've restarted the job several times and it either just fails on the script or produces a 502 error with warnings about categorization --- which shouldn't occur.
I'm thinking it's flakiness but not sure.
Can't reproduce locally. | infrastructure | build fails on error this error started today and i m not sure what the problem is the other parallel jobs worked warning don t know how to categorize amp subscriptions google md skipping home travis build ampproject docs scripts import docs js throw err error error at childprocess exithandler child process js at emittwo events js at childprocess emit events js at maybeclose internal child process js at socket stream socket on internal child process js at emitone events js at socket emit events js at pipe handle close net js build errored after s the following tasks did not complete optimize images sass did you forget to signal async completion log i ve restarted the job several times and it either just fails on the script or produces a error with warnings about categorization which shouldn t occur i m thinking it s flakiness but not sure can t reproduce locally | 1 |
26,967 | 20,969,088,494 | IssuesEvent | 2022-03-28 09:38:53 | tempus-finance/tempus-app | https://api.github.com/repos/tempus-finance/tempus-app | opened | Setup project base for v3 | enhancement infrastructure | that may include but not limited to:
- start with CRA
- pick tech stack, e.g. `@hookstate`, i18n libraries, wallet dependencies, etc
- set up `react-router`
- folder strcuture
- anything you can think of | 1.0 | Setup project base for v3 - that may include but not limited to:
- start with CRA
- pick tech stack, e.g. `@hookstate`, i18n libraries, wallet dependencies, etc
- set up `react-router`
- folder strcuture
- anything you can think of | infrastructure | setup project base for that may include but not limited to start with cra pick tech stack e g hookstate libraries wallet dependencies etc set up react router folder strcuture anything you can think of | 1 |
21,473 | 14,589,107,822 | IssuesEvent | 2020-12-19 00:32:15 | KaTeX/KaTeX | https://api.github.com/repos/KaTeX/KaTeX | closed | Safari screenshotter | bug infrastructure | The Safari screenshotter is broken. I have three PRs that currently are shown as having errors purely due to Safari screenshotter disfunction. Even when the Safari screenshotter works, the process is clumsy and tedious.
I think we should get rid of it. | 1.0 | Safari screenshotter - The Safari screenshotter is broken. I have three PRs that currently are shown as having errors purely due to Safari screenshotter disfunction. Even when the Safari screenshotter works, the process is clumsy and tedious.
I think we should get rid of it. | infrastructure | safari screenshotter the safari screenshotter is broken i have three prs that currently are shown as having errors purely due to safari screenshotter disfunction even when the safari screenshotter works the process is clumsy and tedious i think we should get rid of it | 1 |
3,425 | 4,311,296,346 | IssuesEvent | 2016-07-21 22:09:41 | dart-lang/www.dartlang.org | https://api.github.com/repos/dart-lang/www.dartlang.org | closed | Highlighted TOC section doesn't quite match the currently visible section | bug Infrastructure P2 PatchesWelcome | From https://code.google.com/p/dart/issues/detail?id=21272:
> Hello team :)
>
> https://www.dartlang.org/tools/editor/
>
> Your topic selection list on the left does not work correctly.
>
> All but the first and last selections do not change their own font to bold when selected.
>
> I believe this happens when at certain scroll points.
>
> I am using Google Chrome.
>
> Kindest Regards,
>
> Daniel Couillard
| 1.0 | Highlighted TOC section doesn't quite match the currently visible section - From https://code.google.com/p/dart/issues/detail?id=21272:
> Hello team :)
>
> https://www.dartlang.org/tools/editor/
>
> Your topic selection list on the left does not work correctly.
>
> All but the first and last selections do not change their own font to bold when selected.
>
> I believe this happens when at certain scroll points.
>
> I am using Google Chrome.
>
> Kindest Regards,
>
> Daniel Couillard
| infrastructure | highlighted toc section doesn t quite match the currently visible section from hello team your topic selection list on the left does not work correctly all but the first and last selections do not change their own font to bold when selected i believe this happens when at certain scroll points i am using google chrome kindest regards daniel couillard | 1 |
35,124 | 30,774,216,654 | IssuesEvent | 2023-07-31 04:26:37 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Code Analyzer: Use of wp-env is problematic | status: prioritization tool: monorepo infrastructure | <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
When running Code Analyzer we run into a lot of issues.
```
Command failed: wp-env start
```
wp-env is responsible for calling two functions vi WP CLI
https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/src/Internal/DataStores/Orders/OrdersTableDataStore.php#L2321
https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/bin/wc-get-schema.php#L12-L20
Calling `wp-env start` programatically appears to cause problems, especially with Docker. We need to either:
1. Find a safer way of handling wp-env
2. Find a replacement for wp-env
3. Determine a different way of detecting schema changes, perhaps through static analysis
| 1.0 | Code Analyzer: Use of wp-env is problematic - <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
When running Code Analyzer we run into a lot of issues.
```
Command failed: wp-env start
```
wp-env is responsible for calling two functions vi WP CLI
https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/src/Internal/DataStores/Orders/OrdersTableDataStore.php#L2321
https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/bin/wc-get-schema.php#L12-L20
Calling `wp-env start` programatically appears to cause problems, especially with Docker. We need to either:
1. Find a safer way of handling wp-env
2. Find a replacement for wp-env
3. Determine a different way of detecting schema changes, perhaps through static analysis
| infrastructure | code analyzer use of wp env is problematic prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description when running code analyzer we run into a lot of issues command failed wp env start wp env is responsible for calling two functions vi wp cli calling wp env start programatically appears to cause problems especially with docker we need to either find a safer way of handling wp env find a replacement for wp env determine a different way of detecting schema changes perhaps through static analysis | 1 |
24,799 | 17,786,452,842 | IssuesEvent | 2021-08-31 11:40:57 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Serve built static files in production | type: bug work: infrastructure status: ready | ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
Currently, the frontend is only set up for local development and depends on the Vite server running at port 3000 to serve the frontend. This will not work for production, since we'll have pre-built static files.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We need to update the Django templates to use the existing setup for development environments and serve pre-built static files for production environments (or if the vite server is not running).
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
This is required for https://github.com/centerofci/mathesar/issues/225 | 1.0 | Serve built static files in production - ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
Currently, the frontend is only set up for local development and depends on the Vite server running at port 3000 to serve the frontend. This will not work for production, since we'll have pre-built static files.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We need to update the Django templates to use the existing setup for development environments and serve pre-built static files for production environments (or if the vite server is not running).
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
This is required for https://github.com/centerofci/mathesar/issues/225 | infrastructure | serve built static files in production problem currently the frontend is only set up for local development and depends on the vite server running at port to serve the frontend this will not work for production since we ll have pre built static files proposed solution we need to update the django templates to use the existing setup for development environments and serve pre built static files for production environments or if the vite server is not running additional context this is required for | 1 |
29,733 | 24,222,510,132 | IssuesEvent | 2022-09-26 12:04:45 | arduino/arduino-ide | https://api.github.com/repos/arduino/arduino-ide | closed | Spurious IDE update notifications when using snapshot or nightly builds | topic: infrastructure type: imperfection | ### Describe the problem
On every startup, the Arduino IDE checks for new versions of the IDE (https://github.com/arduino/arduino-ide/pull/797). If a newer version is available, a notification/dialog is shown offering an update.
"Newer" is determined by comparing the version of the user's IDE to the latest available version on the update channel. This comparison is done according to [semver](https://semver.org/).
In order to facilitate beta testing, builds are generated of the Arduino IDE at the current stage in development. These builds are given an identifying version of the following form:
- `<version>-snapshot-<short hash>` - builds generated for every push and pull request that modifies relevant files
- `<version>-nightly-<YYYYMMDD>` - daily builds of the tip of the default branch
The established project management practices cause the `<version>` component of these to be the version of the most recent release.
During the pre-release phase of the project development, all releases had [a pre-release suffix](https://semver.org/#spec-item-9) (e.g., `2.0.0-rc9.4`). Appending the "snapshot" or "nightly" suffix to that pre-release version caused these builds to have the correct precedence (e.g., `2.0.0-rc9.2.snapshot-20cc34c` > `2.0.0-rc9.2`). This situation has changed now that the project is using production release versions (e.g., `2.0.0-nightly-20220915` < `2.0.0`)
🐛 Users of "snapshot" or "nightly" builds are presented with a spurious update notification on startup.
### To reproduce
#### Setup
Make sure you have the default values of the relevant settings:
```json
{
"arduino.checkForUpdates": true,
"arduino.ide.updateChannel": "stable"
}
```
#### Steps
1. Download the latest nightly build of Arduino IDE 2.x from the links on the "Software" page:
https://www.arduino.cc/en/software#nightly-builds
1. Start the nightly build.
🐛 A spurious IDE update notification dialog appears:

### Expected behavior
IDE update notification is only shown when a newer version of the Arduino IDE is available.
### Arduino IDE version
2.0.0-snapshot-5f82577
### Operating system
Windows
### Operating system version
10
### Additional context
I think the solution will be to add a step to the end of the [release procedure](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md) to bump the version metadata after completing a release, following the same procedure as documented under [the "Create the release on GitHub" section of that document](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md#-create-the-release-on-github).
I see two possible approaches:
- Use a pre-release of a patch bump (e.g., the metadata should currently be set to `2.0.1-rc.1`)
- Use a patch bump (e.g., the metadata should currently be set to `2.0.1`)
With the latter approach, the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump (according to the nature of the changes being shipped).
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details
| 1.0 | Spurious IDE update notifications when using snapshot or nightly builds - ### Describe the problem
On every startup, the Arduino IDE checks for new versions of the IDE (https://github.com/arduino/arduino-ide/pull/797). If a newer version is available, a notification/dialog is shown offering an update.
"Newer" is determined by comparing the version of the user's IDE to the latest available version on the update channel. This comparison is done according to [semver](https://semver.org/).
In order to facilitate beta testing, builds are generated of the Arduino IDE at the current stage in development. These builds are given an identifying version of the following form:
- `<version>-snapshot-<short hash>` - builds generated for every push and pull request that modifies relevant files
- `<version>-nightly-<YYYYMMDD>` - daily builds of the tip of the default branch
The established project management practices cause the `<version>` component of these to be the version of the most recent release.
During the pre-release phase of the project development, all releases had [a pre-release suffix](https://semver.org/#spec-item-9) (e.g., `2.0.0-rc9.4`). Appending the "snapshot" or "nightly" suffix to that pre-release version caused these builds to have the correct precedence (e.g., `2.0.0-rc9.2.snapshot-20cc34c` > `2.0.0-rc9.2`). This situation has changed now that the project is using production release versions (e.g., `2.0.0-nightly-20220915` < `2.0.0`)
🐛 Users of "snapshot" or "nightly" builds are presented with a spurious update notification on startup.
### To reproduce
#### Setup
Make sure you have the default values of the relevant settings:
```json
{
"arduino.checkForUpdates": true,
"arduino.ide.updateChannel": "stable"
}
```
#### Steps
1. Download the latest nightly build of Arduino IDE 2.x from the links on the "Software" page:
https://www.arduino.cc/en/software#nightly-builds
1. Start the nightly build.
🐛 A spurious IDE update notification dialog appears:

### Expected behavior
IDE update notification is only shown when a newer version of the Arduino IDE is available.
### Arduino IDE version
2.0.0-snapshot-5f82577
### Operating system
Windows
### Operating system version
10
### Additional context
I think the solution will be to add a step to the end of the [release procedure](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md) to bump the version metadata after completing a release, following the same procedure as documented under [the "Create the release on GitHub" section of that document](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md#-create-the-release-on-github).
I see two possible approaches:
- Use a pre-release of a patch bump (e.g., the metadata should currently be set to `2.0.1-rc.1`)
- Use a patch bump (e.g., the metadata should currently be set to `2.0.1`)
With the latter approach, the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump (according to the nature of the changes being shipped).
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details
| infrastructure | spurious ide update notifications when using snapshot or nightly builds describe the problem on every startup the arduino ide checks for new versions of the ide if a newer version is available a notification dialog is shown offering an update newer is determined by comparing the version of the user s ide to the latest available version on the update channel this comparison is done according to in order to facilitate beta testing builds are generated of the arduino ide at the current stage in development these builds are given an identifying version of the following form snapshot builds generated for every push and pull request that modifies relevant files nightly daily builds of the tip of the default branch the established project management practices cause the component of these to be the version of the most recent release during the pre release phase of the project development all releases had e g appending the snapshot or nightly suffix to that pre release version caused these builds to have the correct precedence e g snapshot this situation has changed now that the project is using production release versions e g nightly 🐛 users of snapshot or nightly builds are presented with a spurious update notification on startup to reproduce setup make sure you have the default values of the relevant settings json arduino checkforupdates true arduino ide updatechannel stable steps download the latest nightly build of arduino ide x from the links on the software page start the nightly build 🐛 a spurious ide update notification dialog appears expected behavior ide update notification is only shown when a newer version of the arduino ide is available arduino ide version snapshot operating system windows operating system version additional context i think the solution will be to add a step to the end of the to bump the version metadata after completing a release following the same procedure as documented under i see two possible approaches use a pre release of a patch bump e g the metadata should currently be set to rc use a patch bump e g the metadata should currently be set to with the latter approach the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump according to the nature of the changes being shipped issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details | 1 |
114,610 | 9,744,504,814 | IssuesEvent | 2019-06-03 07:10:40 | ITISFoundation/osparc-simcore | https://api.github.com/repos/ITISFoundation/osparc-simcore | closed | ISAN Use Case UCD 1D, 2D: Movie Not Shown | ISAN Prio 1 Bug UX_improvements tester_review | ## Long story short
Enter CC 2D and open the cc-2d-viewer. Then retrieve and refresh -> the movie should properly show within the Jupyter notebook.
-> The movie is not shown, does not work.
## Expected behaviour
The movie in the Jupyter notebook should properly show.
## Actual behaviour
Please see recording as follows:

## Steps to reproduce
Please see above.
## Your environment
```json
[
{
"name": "osparc-simcore",
"version": "cfc73c1",
"url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/"
},
{
"name": "osparc-simcore UI ['clean']",
"version": "cfc73c1",
"url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/services/web/client/"
},
{
"name": "qooxdoo-compiler",
"version": "0.3.0-alpha.20190423-1234",
"url": "https://github.com/qooxdoo/qooxdoo-compiler"
},
{
"name": "qxapp",
"version": "1.0.0",
"url": ""
},
{
"name": "qooxdoo framework",
"version": "6.0.0-alpha-20190409-f79cc78",
"url": "http://qooxdoo.org"
},
{
"name": "qx-iconfont-fontawesome5",
"version": "0.0.6",
"url": "https://github.com/ITISFoundation/qx-iconfont-fontawesome5"
},
{
"name": "qx-iconfont-material",
"version": "0.1.4",
"url": "https://github.com/ITISFoundation/qx-osparc-theme"
},
{
"name": "qx-osparc-theme",
"version": "0.4.6",
"url": "https://github.com/ITISFoundation/qx-osparc-theme"
},
{
"name": "socket.io",
"version": "2.1.2",
"url": "https://github.com/socketio/socket.io"
},
{
"name": "jsondiffpatch",
"version": "0.3.11",
"url": "https://github.com/benjamine/jsondiffpatch"
},
{
"name": "jsonTreeViewer",
"version": "0.6.0",
"url": "https://github.com/summerstyle/jsonTreeViewer"
},
{
"name": "Ajv",
"version": "6.5.0",
"url": "https://github.com/epoberezkin/ajv"
},
{
"name": "plotly",
"version": "1.43.2",
"url": "https://github.com/plotly/plotly.js"
},
{
"name": "gridster",
"version": "0.7.0",
"url": "https://github.com/dsmorse/gridster.js"
},
{
"name": "svg.js",
"version": "2.6.4",
"url": "https://github.com/svgdotjs/svg.js"
}
]
``` | 1.0 | ISAN Use Case UCD 1D, 2D: Movie Not Shown - ## Long story short
Enter CC 2D and open the cc-2d-viewer. Then retrieve and refresh -> the movie should properly show within the Jupyter notebook.
-> The movie is not shown, does not work.
## Expected behaviour
The movie in the Jupyter notebook should properly show.
## Actual behaviour
Please see recording as follows:

## Steps to reproduce
Please see above.
## Your environment
```json
[
{
"name": "osparc-simcore",
"version": "cfc73c1",
"url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/"
},
{
"name": "osparc-simcore UI ['clean']",
"version": "cfc73c1",
"url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/services/web/client/"
},
{
"name": "qooxdoo-compiler",
"version": "0.3.0-alpha.20190423-1234",
"url": "https://github.com/qooxdoo/qooxdoo-compiler"
},
{
"name": "qxapp",
"version": "1.0.0",
"url": ""
},
{
"name": "qooxdoo framework",
"version": "6.0.0-alpha-20190409-f79cc78",
"url": "http://qooxdoo.org"
},
{
"name": "qx-iconfont-fontawesome5",
"version": "0.0.6",
"url": "https://github.com/ITISFoundation/qx-iconfont-fontawesome5"
},
{
"name": "qx-iconfont-material",
"version": "0.1.4",
"url": "https://github.com/ITISFoundation/qx-osparc-theme"
},
{
"name": "qx-osparc-theme",
"version": "0.4.6",
"url": "https://github.com/ITISFoundation/qx-osparc-theme"
},
{
"name": "socket.io",
"version": "2.1.2",
"url": "https://github.com/socketio/socket.io"
},
{
"name": "jsondiffpatch",
"version": "0.3.11",
"url": "https://github.com/benjamine/jsondiffpatch"
},
{
"name": "jsonTreeViewer",
"version": "0.6.0",
"url": "https://github.com/summerstyle/jsonTreeViewer"
},
{
"name": "Ajv",
"version": "6.5.0",
"url": "https://github.com/epoberezkin/ajv"
},
{
"name": "plotly",
"version": "1.43.2",
"url": "https://github.com/plotly/plotly.js"
},
{
"name": "gridster",
"version": "0.7.0",
"url": "https://github.com/dsmorse/gridster.js"
},
{
"name": "svg.js",
"version": "2.6.4",
"url": "https://github.com/svgdotjs/svg.js"
}
]
``` | non_infrastructure | isan use case ucd movie not shown long story short enter cc and open the cc viewer then retrieve and refresh the movie should properly show within the jupyter notebook the movie is not shown does not work expected behaviour the movie in the jupyter notebook should properly show actual behaviour please see recording as follows steps to reproduce please see above your environment json name osparc simcore version url name osparc simcore ui version url name qooxdoo compiler version alpha url name qxapp version url name qooxdoo framework version alpha url name qx iconfont version url name qx iconfont material version url name qx osparc theme version url name socket io version url name jsondiffpatch version url name jsontreeviewer version url name ajv version url name plotly version url name gridster version url name svg js version url | 0 |
124,861 | 4,934,574,169 | IssuesEvent | 2016-11-28 19:28:29 | devinit/D-Portal | https://api.github.com/repos/devinit/D-Portal | closed | Update homepage text - Search, and notes | priority | 1 - Add text at top introducing d-portal and new Search function
2 - Update text at the bottom of the page | 1.0 | Update homepage text - Search, and notes - 1 - Add text at top introducing d-portal and new Search function
2 - Update text at the bottom of the page | non_infrastructure | update homepage text search and notes add text at top introducing d portal and new search function update text at the bottom of the page | 0 |
15,309 | 11,455,432,015 | IssuesEvent | 2020-02-06 19:05:20 | enarx/enarx | https://api.github.com/repos/enarx/enarx | closed | Ensure all files have up-to-date Apache license headers | infrastructure | With the new year and all the new project files that come with it, we should ensure everything contains proper license headers. | 1.0 | Ensure all files have up-to-date Apache license headers - With the new year and all the new project files that come with it, we should ensure everything contains proper license headers. | infrastructure | ensure all files have up to date apache license headers with the new year and all the new project files that come with it we should ensure everything contains proper license headers | 1 |
468,467 | 13,483,450,860 | IssuesEvent | 2020-09-11 03:57:02 | Apexal/late | https://api.github.com/repos/Apexal/late | closed | Calendar sometimes duplicates work blocks | Priority: High Type: Bug | When manually adding a work block on the calendar, it sometimes displays two matching blocks after requesting one to be added (ex. two "Homework 3" blocks for the same class within the same time period). It doesn't go away- it has to be manually deleted.
To duplicate:
1. Add work block to calendar.
2. See if duplicate assignment appears.
Device: Windows PC using Google Chrome
| 1.0 | Calendar sometimes duplicates work blocks - When manually adding a work block on the calendar, it sometimes displays two matching blocks after requesting one to be added (ex. two "Homework 3" blocks for the same class within the same time period). It doesn't go away- it has to be manually deleted.
To duplicate:
1. Add work block to calendar.
2. See if duplicate assignment appears.
Device: Windows PC using Google Chrome
| non_infrastructure | calendar sometimes duplicates work blocks when manually adding a work block on the calendar it sometimes displays two matching blocks after requesting one to be added ex two homework blocks for the same class within the same time period it doesn t go away it has to be manually deleted to duplicate add work block to calendar see if duplicate assignment appears device windows pc using google chrome | 0 |
71,248 | 15,190,323,461 | IssuesEvent | 2021-02-15 17:44:59 | gigantum/gigantum-client | https://api.github.com/repos/gigantum/gigantum-client | opened | Add support for user defined HTTPS certs | enhancement feedback:request security team:backend | ## Description
With the improved authentication workflow and future Team Server improvements, it should be easy to run a multi-tenant Client as an externally hosted service with proper TLS certificates. An initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user. Future improvements could automatically run Let's Encrypt certbot. | True | Add support for user defined HTTPS certs - ## Description
With the improved authentication workflow and future Team Server improvements, it should be easy to run a multi-tenant Client as an externally hosted service with proper TLS certificates. An initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user. Future improvements could automatically run Let's Encrypt certbot. | non_infrastructure | add support for user defined https certs description with the improved authentication workflow and future team server improvements it should be easy to run a multi tenant client as an externally hosted service with proper tls certificates an initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user future improvements could automatically run let s encrypt certbot | 0 |
252,748 | 21,628,987,244 | IssuesEvent | 2022-05-05 07:40:56 | storj/gateway-mt | https://api.github.com/repos/storj/gateway-mt | closed | Custom metadata is not saved when S3 MPU is used | bug s3-tests blocked | Custom metadata is not saved when S3 MPU is used (in this example the threshold is 8MB)
As you can see in the second case the metadata is empty, but with real AWS it's foo: bar in both cases.
```
> aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 1mb s3://bucket1/1mb
upload: ./1mb to s3://bucket1/1mb
> aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 20mb s3://bucket1/20mb
upload: ./20mb to s3://bucket1/20mb
> aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 1mb x
{
"AcceptRanges": "bytes",
"LastModified": "Thu, 30 Sep 2021 11:36:10 GMT",
"ContentLength": 1048576,
"ETag": "\"f5f16eee8ad5a86a3439dc75429831ee\"",
"ContentType": "application/octet-stream",
"Metadata": {
"foo": "bar"
}
}
> aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 20mb x
{
"AcceptRanges": "bytes",
"LastModified": "Thu, 30 Sep 2021 11:36:21 GMT",
"ContentLength": 20971520,
"ETag": "\"80f096bdc4c7fd9639307cd1174b398a-3\"",
"ContentType": "application/octet-stream",
"Metadata": {}
}
```
Credit:
The problem is identified by the rclone s3 unit tests during the conversation in rclone/rclone#5616 where @olefrost suggested to re-run s3 unit tests with storj gateway-mt and @ncw drew attention to this problem from many failures. | 1.0 | Custom metadata is not saved when S3 MPU is used - Custom metadata is not saved when S3 MPU is used (in this example the threshold is 8MB)
As you can see in the second case the metadata is empty, but with real AWS it's foo: bar in both cases.
```
> aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 1mb s3://bucket1/1mb
upload: ./1mb to s3://bucket1/1mb
> aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 20mb s3://bucket1/20mb
upload: ./20mb to s3://bucket1/20mb
> aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 1mb x
{
"AcceptRanges": "bytes",
"LastModified": "Thu, 30 Sep 2021 11:36:10 GMT",
"ContentLength": 1048576,
"ETag": "\"f5f16eee8ad5a86a3439dc75429831ee\"",
"ContentType": "application/octet-stream",
"Metadata": {
"foo": "bar"
}
}
> aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 20mb x
{
"AcceptRanges": "bytes",
"LastModified": "Thu, 30 Sep 2021 11:36:21 GMT",
"ContentLength": 20971520,
"ETag": "\"80f096bdc4c7fd9639307cd1174b398a-3\"",
"ContentType": "application/octet-stream",
"Metadata": {}
}
```
Credit:
The problem is identified by the rclone s3 unit tests during the conversation in rclone/rclone#5616 where @olefrost suggested to re-run s3 unit tests with storj gateway-mt and @ncw drew attention to this problem from many failures. | non_infrastructure | custom metadata is not saved when mpu is used custom metadata is not saved when mpu is used in this example the threshold is as you can see in the second case the metadata is empty but with real aws it s foo bar in both cases aws cp endpoint metadata foo bar upload to aws cp endpoint metadata foo bar upload to aws get object endpoint bucket key x acceptranges bytes lastmodified thu sep gmt contentlength etag contenttype application octet stream metadata foo bar aws get object endpoint bucket key x acceptranges bytes lastmodified thu sep gmt contentlength etag contenttype application octet stream metadata credit the problem is identified by the rclone unit tests during the conversation in rclone rclone where olefrost suggested to re run unit tests with storj gateway mt and ncw drew attention to this problem from many failures | 0 |
12,313 | 9,692,739,967 | IssuesEvent | 2019-05-24 14:28:21 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | Symbols from daily builds of master brancn are not published to symbol server | 2 - Working area-infrastructure question | ## Summary
The C# extension test team recently did a test of symbol server support using a preview6 build of the framework and found that one assembly, Microsoft.AspNetCore.Routing.Abstractions, had a PDB file, but the PDB wasn't published to symbol server.
## Environment data
`dotnet --info` output:
.NET Core SDK (reflecting any global.json):
Version: 3.0.100-preview6-011799
Commit: 3816213dde
Runtime Environment:
OS Name: Mac OS X
OS Version: 10.14
OS Platform: Darwin
RID: osx.10.14-x64
Base Path: /usr/local/share/dotnet/sdk/3.0.100-preview6-011799/
Host (useful for support):
Version: 3.0.0-preview6-27707-07
Commit: c9e7f134dc
.NET Core SDKs installed:
3.0.100-preview6-011799 [/usr/local/share/dotnet/sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.App 3.0.0-preview6-19257-04 [/usr/local/share/dotnet/shared/Microsoft.AspNetCore.App]
Microsoft.NETCore.App 3.0.0-preview6-27707-07 [/usr/local/share/dotnet/shared/Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
VS Code version: 1.33.1
C# Extension version: 1.19 (v1.19.1)
## Steps to reproduce
1. Create a new ASP.NET Razor project by doing
mkdir razor
cd razor
dotnet new razor
2. Open razor with VScode
3. It may ask at the bottom right to install assets to build and debug, say yes.
4. After it has done this it will have created a .vscode directory in the ‘razor’ directory and a ‘launch.json’ file in that directory. Open the ‘.vscode’ directory, and double click on the ‘launch.json to edit it.
5. Add the following lines after the “name” line of the “Configurations” object. (This allows debugging through the framework.
"justMyCode": false,
"symbolOptions": {
"searchMicrosoftSymbolServer": true
},
"suppressJITOptimizations": true,
6. Double click on the ‘Startup.cs’ file in the ‘Explorer’ pane to bring up that source code.
7. Add the following line to the Startup.cs file at the beginning of the ‘Configure’ method (at the bottom of the file).
Console.WriteLine("x " + 3);
8. Place a breakpoint on that line (F9).
9. Run the application until it hit that breakpoint (F5).
## Expected behavior
It will load symbols from the Microsoft symbol server successfully.
## Actual behavior
One assembly, Microsoft.AspNetCore.Routing.Abstractions, didn't have symbols. | 1.0 | Symbols from daily builds of master brancn are not published to symbol server - ## Summary
The C# extension test team recently did a test of symbol server support using a preview6 build of the framework and found that one assembly, Microsoft.AspNetCore.Routing.Abstractions, had a PDB file, but the PDB wasn't published to symbol server.
## Environment data
`dotnet --info` output:
.NET Core SDK (reflecting any global.json):
Version: 3.0.100-preview6-011799
Commit: 3816213dde
Runtime Environment:
OS Name: Mac OS X
OS Version: 10.14
OS Platform: Darwin
RID: osx.10.14-x64
Base Path: /usr/local/share/dotnet/sdk/3.0.100-preview6-011799/
Host (useful for support):
Version: 3.0.0-preview6-27707-07
Commit: c9e7f134dc
.NET Core SDKs installed:
3.0.100-preview6-011799 [/usr/local/share/dotnet/sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.App 3.0.0-preview6-19257-04 [/usr/local/share/dotnet/shared/Microsoft.AspNetCore.App]
Microsoft.NETCore.App 3.0.0-preview6-27707-07 [/usr/local/share/dotnet/shared/Microsoft.NETCore.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
VS Code version: 1.33.1
C# Extension version: 1.19 (v1.19.1)
## Steps to reproduce
1. Create a new ASP.NET Razor project by doing
mkdir razor
cd razor
dotnet new razor
2. Open razor with VScode
3. It may ask at the bottom right to install assets to build and debug, say yes.
4. After it has done this it will have created a .vscode directory in the ‘razor’ directory and a ‘launch.json’ file in that directory. Open the ‘.vscode’ directory, and double click on the ‘launch.json to edit it.
5. Add the following lines after the “name” line of the “Configurations” object. (This allows debugging through the framework.
"justMyCode": false,
"symbolOptions": {
"searchMicrosoftSymbolServer": true
},
"suppressJITOptimizations": true,
6. Double click on the ‘Startup.cs’ file in the ‘Explorer’ pane to bring up that source code.
7. Add the following line to the Startup.cs file at the beginning of the ‘Configure’ method (at the bottom of the file).
Console.WriteLine("x " + 3);
8. Place a breakpoint on that line (F9).
9. Run the application until it hit that breakpoint (F5).
## Expected behavior
It will load symbols from the Microsoft symbol server successfully.
## Actual behavior
One assembly, Microsoft.AspNetCore.Routing.Abstractions, didn't have symbols. | infrastructure | symbols from daily builds of master brancn are not published to symbol server summary the c extension test team recently did a test of symbol server support using a build of the framework and found that one assembly microsoft aspnetcore routing abstractions had a pdb file but the pdb wasn t published to symbol server environment data dotnet info output net core sdk reflecting any global json version commit runtime environment os name mac os x os version os platform darwin rid osx base path usr local share dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore app microsoft netcore app to install additional net core runtimes or sdks vs code version c extension version steps to reproduce create a new asp net razor project by doing mkdir razor cd razor dotnet new razor open razor with vscode it may ask at the bottom right to install assets to build and debug say yes after it has done this it will have created a vscode directory in the ‘razor’ directory and a ‘launch json’ file in that directory open the ‘ vscode’ directory and double click on the ‘launch json to edit it add the following lines after the “name” line of the “configurations” object this allows debugging through the framework justmycode false symboloptions searchmicrosoftsymbolserver true suppressjitoptimizations true double click on the ‘startup cs’ file in the ‘explorer’ pane to bring up that source code add the following line to the startup cs file at the beginning of the ‘configure’ method at the bottom of the file console writeline x place a breakpoint on that line run the application until it hit that breakpoint expected behavior it will load symbols from the microsoft symbol server successfully actual behavior one assembly microsoft aspnetcore routing abstractions didn t have symbols | 1 |
107,014 | 16,743,776,674 | IssuesEvent | 2021-06-11 13:12:33 | Polkadex-Substrate/polkadexTEE-worker | https://api.github.com/repos/Polkadex-Substrate/polkadexTEE-worker | opened | Exchange websocket client with rustls client | security | Exchange websocket client with rustls client as soon as openfinex supports tls | True | Exchange websocket client with rustls client - Exchange websocket client with rustls client as soon as openfinex supports tls | non_infrastructure | exchange websocket client with rustls client exchange websocket client with rustls client as soon as openfinex supports tls | 0 |
26,489 | 20,156,851,176 | IssuesEvent | 2022-02-09 17:13:49 | SasView/sasview | https://api.github.com/repos/SasView/sasview | opened | Decide what to do about the SourceForge and danse source code archives and the archive repo | Admin and support infrastructure task | _This issue has been written as an aide memoir for @butlerpd and @smk78!_
A long-standing issue has been to sunset the old SANSView/SasView SourceForge site. A pre-requisite for this is ensuring important content is securely archived elsewhere, and GitHub would be the logical place.
A similar, but arguably less urgent, argument applies to legacy content on the danse server.
The SasView\archive repo was created for this purpose, and is functional. However, a problem has arisen.
It transpires that GitHub will not track files >~50Mb, or >~25Mb if using the browser interface. For such files, GitHub recommends doing one of two things:
- either, adding these files using [Git Large File Support (Git-LFS)](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-git-large-file-storage),
- or, by adding these files as releases.
For simplicity, @smk78 migrated the legacy .gz, .tgz, and .zip distribution archives to \archive using Git-LFS.
This has worked well, except the volume of data has slightly exceeded (1.06Gb) the free 1Gb LFS limit that GitHub makes available per organization.
There is no immediate threat to the content in \archive (see [here](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-storage-and-bandwidth-usage)) and the repo can still be cloned, but no additional content _anywhere_ in the SasView organization can now be added by LFS **(adding by normal commits is unaffected!)**.
To add further content by LFS would require removing existing content, or purchasing a monthly Git-LFS 'data pack' (at the time of writing $5 gets 50Gb).
Alternatively, we need to figure out another way to archive our legacy content.
| 1.0 | Decide what to do about the SourceForge and danse source code archives and the archive repo - _This issue has been written as an aide memoir for @butlerpd and @smk78!_
A long-standing issue has been to sunset the old SANSView/SasView SourceForge site. A pre-requisite for this is ensuring important content is securely archived elsewhere, and GitHub would be the logical place.
A similar, but arguably less urgent, argument applies to legacy content on the danse server.
The SasView\archive repo was created for this purpose, and is functional. However, a problem has arisen.
It transpires that GitHub will not track files >~50Mb, or >~25Mb if using the browser interface. For such files, GitHub recommends doing one of two things:
- either, adding these files using [Git Large File Support (Git-LFS)](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-git-large-file-storage),
- or, by adding these files as releases.
For simplicity, @smk78 migrated the legacy .gz, .tgz, and .zip distribution archives to \archive using Git-LFS.
This has worked well, except the volume of data has slightly exceeded (1.06Gb) the free 1Gb LFS limit that GitHub makes available per organization.
There is no immediate threat to the content in \archive (see [here](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-storage-and-bandwidth-usage)) and the repo can still be cloned, but no additional content _anywhere_ in the SasView organization can now be added by LFS **(adding by normal commits is unaffected!)**.
To add further content by LFS would require removing existing content, or purchasing a monthly Git-LFS 'data pack' (at the time of writing $5 gets 50Gb).
Alternatively, we need to figure out another way to archive our legacy content.
| infrastructure | decide what to do about the sourceforge and danse source code archives and the archive repo this issue has been written as an aide memoir for butlerpd and a long standing issue has been to sunset the old sansview sasview sourceforge site a pre requisite for this is ensuring important content is securely archived elsewhere and github would be the logical place a similar but arguably less urgent argument applies to legacy content on the danse server the sasview archive repo was created for this purpose and is functional however a problem has arisen it transpires that github will not track files or if using the browser interface for such files github recommends doing one of two things either adding these files using or by adding these files as releases for simplicity migrated the legacy gz tgz and zip distribution archives to archive using git lfs this has worked well except the volume of data has slightly exceeded the free lfs limit that github makes available per organization there is no immediate threat to the content in archive see and the repo can still be cloned but no additional content anywhere in the sasview organization can now be added by lfs adding by normal commits is unaffected to add further content by lfs would require removing existing content or purchasing a monthly git lfs data pack at the time of writing gets alternatively we need to figure out another way to archive our legacy content | 1 |
24,398 | 17,196,122,245 | IssuesEvent | 2021-07-16 17:39:26 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | opened | Network layer changes for hooking up the backend for PlatformParameters | Priority: Essential Type: Task Where: Infrastructure | PlatformParameter API & Gae Models, also a MockPlatformParameterSerivice for fake response.
PlatformParameterService
GaePlatformParameters
GaePlatformParameter
MockPlatformParameterService
Target PR date: 17 July 2021
Target completion date: 24 July 2021 | 1.0 | Network layer changes for hooking up the backend for PlatformParameters - PlatformParameter API & Gae Models, also a MockPlatformParameterSerivice for fake response.
PlatformParameterService
GaePlatformParameters
GaePlatformParameter
MockPlatformParameterService
Target PR date: 17 July 2021
Target completion date: 24 July 2021 | infrastructure | network layer changes for hooking up the backend for platformparameters platformparameter api gae models also a mockplatformparameterserivice for fake response platformparameterservice gaeplatformparameters gaeplatformparameter mockplatformparameterservice target pr date july target completion date july | 1 |
466,893 | 13,436,452,446 | IssuesEvent | 2020-09-07 14:24:13 | scikit-image/scikit-image | https://api.github.com/repos/scikit-image/scikit-image | closed | Memory corruption in flood_fill | action: solution identified good first issue priority: critical type: bug | ## Description
When using `flood_fill`, there is a consistent double free error and a dying kernel (as shown in the output from Jupyter) when using a negative row coordinate for the point where the fill starts.
## Way to reproduce
```python
import numpy as np
from skimage import segmentation
a = np.random.randint(0, 2, (100,100))
#for r in range(a.shape[0]):
# segmentation.flood_fill(a, (r, 0), 1, in_place=True)
# segmentation.flood_fill(a, (r, a.shape[1]-1), 1, in_place=True)
# segmentation.flood_fill(a, (r, -1), 1, in_place=True)
for c in range(a.shape[1]):
# segmentation.flood_fill(a, ( 0, c), 1, in_place=True)
# segmentation.flood_fill(a, (a.shape[0]-1, c), 1, in_place=True)
segmentation.flood_fill(a, (-1, c), 1, in_place=True)
```
All commented lines work, except the last one, which consistently triggers the error. The `flood` function also fails likewise with `segmentation.flood(a, (-1, c))`.
## Version information
```python
# Paste the output of the following python commands
from __future__ import print_function
import sys; print(sys.version)
import platform; print(platform.platform())
import skimage; print("scikit-image version: {}".format(skimage.__version__))
import numpy; print("numpy version: {}".format(numpy.__version__))
```
```python
3.7.7 (default, May 7 2020, 21:25:33)
[GCC 7.3.0]
Linux-5.4.0-42-generic-x86_64-with-debian-bullseye-sid
scikit-image version: 0.17.2
numpy version: 1.19.1
```
This error was present in version 0.16.2 too.
| 1.0 | Memory corruption in flood_fill - ## Description
When using `flood_fill`, there is a consistent double free error and a dying kernel (as shown in the output from Jupyter) when using a negative row coordinate for the point where the fill starts.
## Way to reproduce
```python
import numpy as np
from skimage import segmentation
a = np.random.randint(0, 2, (100,100))
#for r in range(a.shape[0]):
# segmentation.flood_fill(a, (r, 0), 1, in_place=True)
# segmentation.flood_fill(a, (r, a.shape[1]-1), 1, in_place=True)
# segmentation.flood_fill(a, (r, -1), 1, in_place=True)
for c in range(a.shape[1]):
# segmentation.flood_fill(a, ( 0, c), 1, in_place=True)
# segmentation.flood_fill(a, (a.shape[0]-1, c), 1, in_place=True)
segmentation.flood_fill(a, (-1, c), 1, in_place=True)
```
All commented lines work, except the last one, which consistently triggers the error. The `flood` function also fails likewise with `segmentation.flood(a, (-1, c))`.
## Version information
```python
# Paste the output of the following python commands
from __future__ import print_function
import sys; print(sys.version)
import platform; print(platform.platform())
import skimage; print("scikit-image version: {}".format(skimage.__version__))
import numpy; print("numpy version: {}".format(numpy.__version__))
```
```python
3.7.7 (default, May 7 2020, 21:25:33)
[GCC 7.3.0]
Linux-5.4.0-42-generic-x86_64-with-debian-bullseye-sid
scikit-image version: 0.17.2
numpy version: 1.19.1
```
This error was present in version 0.16.2 too.
| non_infrastructure | memory corruption in flood fill description when using flood fill there is a consistent double free error and a dying kernel as shown in the output from jupyter when using a negative row coordinate for the point where the fill starts way to reproduce python import numpy as np from skimage import segmentation a np random randint for r in range a shape segmentation flood fill a r in place true segmentation flood fill a r a shape in place true segmentation flood fill a r in place true for c in range a shape segmentation flood fill a c in place true segmentation flood fill a a shape c in place true segmentation flood fill a c in place true all commented lines work except the last one which consistently triggers the error the flood function also fails likewise with segmentation flood a c version information python paste the output of the following python commands from future import print function import sys print sys version import platform print platform platform import skimage print scikit image version format skimage version import numpy print numpy version format numpy version python default may linux generic with debian bullseye sid scikit image version numpy version this error was present in version too | 0 |
61,620 | 17,023,741,078 | IssuesEvent | 2021-07-03 03:35:35 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Wrong parent region in Austria | Component: nominatim Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 11.36am, Friday, 19th August 2011]**
When searching adresses in two administrative boundaries (Schrding and Ried) the parent region indicated is "Vcklabruck" which has type: place: region;
This is wrong; all three are administrative entities (counties) on the same level with parent "Obersterreich: boundaries: administrative | 1.0 | Wrong parent region in Austria - **[Submitted to the original trac issue database at 11.36am, Friday, 19th August 2011]**
When searching adresses in two administrative boundaries (Schrding and Ried) the parent region indicated is "Vcklabruck" which has type: place: region;
This is wrong; all three are administrative entities (counties) on the same level with parent "Obersterreich: boundaries: administrative | non_infrastructure | wrong parent region in austria when searching adresses in two administrative boundaries schrding and ried the parent region indicated is vcklabruck which has type place region this is wrong all three are administrative entities counties on the same level with parent obersterreich boundaries administrative | 0 |
309,210 | 26,656,594,690 | IssuesEvent | 2023-01-25 17:18:58 | getsentry/sentry-javascript | https://api.github.com/repos/getsentry/sentry-javascript | closed | Add integration tests for NextJS dev server | Package: Nextjs Status: Backlog Type: Tests | We currently only run integration test for our nextjs sdk against production builds, but we should we should also run integration tests against the dev server for nextjs. We want to make sure our future changes don't break dev workflows.
https://github.com/getsentry/sentry-javascript/issues/4115 was an issue that was not caught on release because it affected only dev server env. | 1.0 | Add integration tests for NextJS dev server - We currently only run integration test for our nextjs sdk against production builds, but we should we should also run integration tests against the dev server for nextjs. We want to make sure our future changes don't break dev workflows.
https://github.com/getsentry/sentry-javascript/issues/4115 was an issue that was not caught on release because it affected only dev server env. | non_infrastructure | add integration tests for nextjs dev server we currently only run integration test for our nextjs sdk against production builds but we should we should also run integration tests against the dev server for nextjs we want to make sure our future changes don t break dev workflows was an issue that was not caught on release because it affected only dev server env | 0 |
12,169 | 9,608,700,856 | IssuesEvent | 2019-05-12 09:00:33 | eclipse/vorto | https://api.github.com/repos/eclipse/vorto | closed | Upgrade db schema on vorto-dev | Infrastructure Repository | Confirmations:
- AWS RDS DB Schema contains changes for multi - tenancy | 1.0 | Upgrade db schema on vorto-dev - Confirmations:
- AWS RDS DB Schema contains changes for multi - tenancy | infrastructure | upgrade db schema on vorto dev confirmations aws rds db schema contains changes for multi tenancy | 1 |
248,599 | 7,934,299,283 | IssuesEvent | 2018-07-08 17:32:28 | Kheiden/RPi-tankbot | https://api.github.com/repos/Kheiden/RPi-tankbot | closed | Single photo disparity map | Priority 2 core | Depends on #43
I want to create a single photo disparity map image. | 1.0 | Single photo disparity map - Depends on #43
I want to create a single photo disparity map image. | non_infrastructure | single photo disparity map depends on i want to create a single photo disparity map image | 0 |
26,396 | 20,071,797,457 | IssuesEvent | 2022-02-04 08:03:10 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Build failure: "...nativeaot\Runtime\Full\AsmOffsets.cs' could not be found" | area-Infrastructure-coreclr untriaged in-pr | Building with:
```
build.cmd clr.runtime+Clr.NativeAotLibs+clr.corelib+clr.nativecorelib+clr.tools+clr.iltools+clr.alljits+libs -arch x64 -c Debug -librariesConfiguration Release
```
fails with:
```
CSC : error CS2001: Source file 'C:\gh\runtime3\artifacts\obj\\coreclr\windows.x64.Debug\nativeaot\Runtime\Full\AsmOffsets.cs' could not be found. [C:\gh\runtime3\src\coreclr\nativeaot\Test.CoreLib\src\Test.CoreLib.csproj]
```
However, building with:
```
build.cmd clr+libs -arch x64 -c Debug -librariesConfiguration Release
```
succeeds.
This seems to indicate that there's some issue with the Clr.NativeAotLibs subset not building enough.
@MichalStrehovsky ? | 1.0 | Build failure: "...nativeaot\Runtime\Full\AsmOffsets.cs' could not be found" - Building with:
```
build.cmd clr.runtime+Clr.NativeAotLibs+clr.corelib+clr.nativecorelib+clr.tools+clr.iltools+clr.alljits+libs -arch x64 -c Debug -librariesConfiguration Release
```
fails with:
```
CSC : error CS2001: Source file 'C:\gh\runtime3\artifacts\obj\\coreclr\windows.x64.Debug\nativeaot\Runtime\Full\AsmOffsets.cs' could not be found. [C:\gh\runtime3\src\coreclr\nativeaot\Test.CoreLib\src\Test.CoreLib.csproj]
```
However, building with:
```
build.cmd clr+libs -arch x64 -c Debug -librariesConfiguration Release
```
succeeds.
This seems to indicate that there's some issue with the Clr.NativeAotLibs subset not building enough.
@MichalStrehovsky ? | infrastructure | build failure nativeaot runtime full asmoffsets cs could not be found building with build cmd clr runtime clr nativeaotlibs clr corelib clr nativecorelib clr tools clr iltools clr alljits libs arch c debug librariesconfiguration release fails with csc error source file c gh artifacts obj coreclr windows debug nativeaot runtime full asmoffsets cs could not be found however building with build cmd clr libs arch c debug librariesconfiguration release succeeds this seems to indicate that there s some issue with the clr nativeaotlibs subset not building enough michalstrehovsky | 1 |
13,750 | 16,502,892,037 | IssuesEvent | 2021-05-25 15:57:23 | googleapis/google-auth-library-python | https://api.github.com/repos/googleapis/google-auth-library-python | closed | 'testing/constraints-*.txt' not used | priority: p2 type: process | From #757.
In [`noxfile.py`](https://github.com/googleapis/google-auth-library-python/blob/a9234423cb2b69068fc0d30a5a0ee86a599ab8b7/noxfile.py#L87-L89), the constraints are not being used, which is why the one populated file, [`testing/constraints-3.6.txt](https://github.com/googleapis/google-auth-library-python/blob/master/testing/constraints-3.6.txt) is in an invalid state (multiple entries for `rsa`, maybe more issues). | 1.0 | 'testing/constraints-*.txt' not used - From #757.
In [`noxfile.py`](https://github.com/googleapis/google-auth-library-python/blob/a9234423cb2b69068fc0d30a5a0ee86a599ab8b7/noxfile.py#L87-L89), the constraints are not being used, which is why the one populated file, [`testing/constraints-3.6.txt](https://github.com/googleapis/google-auth-library-python/blob/master/testing/constraints-3.6.txt) is in an invalid state (multiple entries for `rsa`, maybe more issues). | non_infrastructure | testing constraints txt not used from in the constraints are not being used which is why the one populated file is in an invalid state multiple entries for rsa maybe more issues | 0 |
6,327 | 6,334,663,947 | IssuesEvent | 2017-07-26 17:08:57 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | Update the publish definition to match what is in release/2.0.0 branch | area-Infrastructure | Need to port https://github.com/dotnet/coreclr/commit/ba86da4a6c821b0f042e105a72fd901c7f7b94c3 and https://github.com/dotnet/coreclr/commit/8ab4fb4260bdf1dafdf4b050bcc01102437f875d to master once the builds become stable. | 1.0 | Update the publish definition to match what is in release/2.0.0 branch - Need to port https://github.com/dotnet/coreclr/commit/ba86da4a6c821b0f042e105a72fd901c7f7b94c3 and https://github.com/dotnet/coreclr/commit/8ab4fb4260bdf1dafdf4b050bcc01102437f875d to master once the builds become stable. | infrastructure | update the publish definition to match what is in release branch need to port and to master once the builds become stable | 1 |
85,825 | 15,755,286,339 | IssuesEvent | 2021-03-31 01:30:36 | ChenLuigi/GitHubScannerBower4 | https://api.github.com/repos/ChenLuigi/GitHubScannerBower4 | opened | WS-2017-0247 (Low) detected in ms-0.7.1.tgz | security vulnerability | ## WS-2017-0247 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.7.1.tgz</b></p></summary>
<p>Tiny ms conversion utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.1.tgz">https://registry.npmjs.org/ms/-/ms-0.7.1.tgz</a></p>
<p>Path to dependency file: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/package.json</p>
<p>Path to vulnerable library: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/node_modules/ms/package.json</p>
<p>
Dependency Hierarchy:
- grunt-sass-0.17.0.tgz (Root Library)
- node-sass-1.2.3.tgz
- mocha-2.5.3.tgz
- debug-2.2.0.tgz
- :x: **ms-0.7.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2017-04-12
<p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p>
<p>Release Date: 2017-04-12</p>
<p>Fix Resolution: 2.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2017-0247 (Low) detected in ms-0.7.1.tgz - ## WS-2017-0247 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.7.1.tgz</b></p></summary>
<p>Tiny ms conversion utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.1.tgz">https://registry.npmjs.org/ms/-/ms-0.7.1.tgz</a></p>
<p>Path to dependency file: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/package.json</p>
<p>Path to vulnerable library: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/node_modules/ms/package.json</p>
<p>
Dependency Hierarchy:
- grunt-sass-0.17.0.tgz (Root Library)
- node-sass-1.2.3.tgz
- mocha-2.5.3.tgz
- debug-2.2.0.tgz
- :x: **ms-0.7.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2017-04-12
<p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p>
<p>Release Date: 2017-04-12</p>
<p>Fix Resolution: 2.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | ws low detected in ms tgz ws low severity vulnerability vulnerable library ms tgz tiny ms conversion utility library home page a href path to dependency file goldenpanel lighter goldenpanel package package json path to vulnerable library goldenpanel lighter goldenpanel package node modules ms package json dependency hierarchy grunt sass tgz root library node sass tgz mocha tgz debug tgz x ms tgz vulnerable library vulnerability details affected versions of this package are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
583 | 2,776,011,819 | IssuesEvent | 2015-05-04 19:19:52 | WP-API/WP-API | https://api.github.com/repos/WP-API/WP-API | closed | Write permission discoverability for endpoints | #Blessed Infrastructure | As a client developer writing an application, I want to have my application discover which endpoints are readable / writable through response values, and without having to know the intricacies of WordPress' capabilities system.
Currently, we can't easily support this use case because capability is determined at run time. We'll need some sort of framework for a `WP_JSON_Controller` to expose permission scope for its endpoints. | 1.0 | Write permission discoverability for endpoints - As a client developer writing an application, I want to have my application discover which endpoints are readable / writable through response values, and without having to know the intricacies of WordPress' capabilities system.
Currently, we can't easily support this use case because capability is determined at run time. We'll need some sort of framework for a `WP_JSON_Controller` to expose permission scope for its endpoints. | infrastructure | write permission discoverability for endpoints as a client developer writing an application i want to have my application discover which endpoints are readable writable through response values and without having to know the intricacies of wordpress capabilities system currently we can t easily support this use case because capability is determined at run time we ll need some sort of framework for a wp json controller to expose permission scope for its endpoints | 1 |
938 | 2,521,170,027 | IssuesEvent | 2015-01-19 12:19:05 | ChrisCummins/msc-thesis | https://api.github.com/repos/ChrisCummins/msc-thesis | closed | Build error on Ubuntu using distro automake install. | Bug Testing & Tooling | Ubuntu distro version of automake is 1.11.3. The `AM_EXTRA_RECURSIVE_TARGETS` macro used in `configure.ac` is undefined in automake < 1.13.
See GNU [Automake 1.13 release notes](http://lwn.net/Articles/531373/). | 1.0 | Build error on Ubuntu using distro automake install. - Ubuntu distro version of automake is 1.11.3. The `AM_EXTRA_RECURSIVE_TARGETS` macro used in `configure.ac` is undefined in automake < 1.13.
See GNU [Automake 1.13 release notes](http://lwn.net/Articles/531373/). | non_infrastructure | build error on ubuntu using distro automake install ubuntu distro version of automake is the am extra recursive targets macro used in configure ac is undefined in automake see gnu | 0 |
8,597 | 7,519,415,515 | IssuesEvent | 2018-04-12 11:32:25 | djavorszky/ddn | https://api.github.com/repos/djavorszky/ddn | closed | Containeraize agents | enhancement infrastructure new feature question | Currently, most connectors are running inside containers, but in a manual fashion, that is, the container is started and then the connector is copied inside and then started manually.
I'd like to have containers that already have these (mostly) set up from the moment someone pulls them and starts them. This would allow us to also do sub-second updates on them.
Will require us writing a Dockerfile for each.
- [x] mysql-55
- [x] mysql-56
- [x] mysql-57
- [x] postgres-94 | 1.0 | Containeraize agents - Currently, most connectors are running inside containers, but in a manual fashion, that is, the container is started and then the connector is copied inside and then started manually.
I'd like to have containers that already have these (mostly) set up from the moment someone pulls them and starts them. This would allow us to also do sub-second updates on them.
Will require us writing a Dockerfile for each.
- [x] mysql-55
- [x] mysql-56
- [x] mysql-57
- [x] postgres-94 | infrastructure | containeraize agents currently most connectors are running inside containers but in a manual fashion that is the container is started and then the connector is copied inside and then started manually i d like to have containers that already have these mostly set up from the moment someone pulls them and starts them this would allow us to also do sub second updates on them will require us writing a dockerfile for each mysql mysql mysql postgres | 1 |
22,948 | 15,691,005,849 | IssuesEvent | 2021-03-25 17:21:07 | Alv-no/alv-website | https://api.github.com/repos/Alv-no/alv-website | closed | Setup deployment pipeline on Azure | infrastructure | - Setup Azure DevOps pipeline inspired [by this](https://github.com/Alv-no/alvtime-monorepo/tree/master/.AzureDevOps/pipelines)
- Use Terraform and Docker | 1.0 | Setup deployment pipeline on Azure - - Setup Azure DevOps pipeline inspired [by this](https://github.com/Alv-no/alvtime-monorepo/tree/master/.AzureDevOps/pipelines)
- Use Terraform and Docker | infrastructure | setup deployment pipeline on azure setup azure devops pipeline inspired use terraform and docker | 1 |
77,423 | 9,565,076,341 | IssuesEvent | 2019-05-05 09:32:34 | gitthermal/thermal | https://api.github.com/repos/gitthermal/thermal | opened | Commit related information to repository settings | Design Front end Git discussion 💡 Feature | ## Description
Add commit related information to the repository settings.
#### For example
- General
- Last open
- Last commit
- User description
- Working copy
- Current Branch
- Status
| 1.0 | Commit related information to repository settings - ## Description
Add commit related information to the repository settings.
#### For example
- General
- Last open
- Last commit
- User description
- Working copy
- Current Branch
- Status
| non_infrastructure | commit related information to repository settings description add commit related information to the repository settings for example general last open last commit user description working copy current branch status | 0 |
71,221 | 23,492,766,477 | IssuesEvent | 2022-08-17 20:33:14 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [Bug :bug:]: ModuleNotFoundError: No module named 'certifi' | C-py I-defect needs-triaging | ### What happened?
Error in [this build](https://github.com/alexmojaki/pyodide-worker-runner/runs/7885917987?check_suite_focus=true).
### How can we reproduce the issue?
```shell
from selenium import webdriver
```
### Relevant log output
```shell
test.py:8: in <module>
from selenium import webdriver
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/__init__.py:18: in <module>
from .firefox.webdriver import WebDriver as Firefox # noqa
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/firefox/webdriver.py:26: in <module>
from selenium.webdriver.remote.webdriver import WebDriver as RemoteWebDriver
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py:52: in <module>
from .remote_connection import RemoteConnection
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/remote_connection.py:25: in <module>
import certifi
E ModuleNotFoundError: No module named 'certifi'
```
### Operating System
Ubuntu
### Selenium version
Python 4.4.1
### What are the browser(s) and version(s) where you see this issue?
None
### What are the browser driver(s) and version(s) where you see this issue?
None
### Are you using Selenium Grid?
_No response_ | 1.0 | [Bug :bug:]: ModuleNotFoundError: No module named 'certifi' - ### What happened?
Error in [this build](https://github.com/alexmojaki/pyodide-worker-runner/runs/7885917987?check_suite_focus=true).
### How can we reproduce the issue?
```shell
from selenium import webdriver
```
### Relevant log output
```shell
test.py:8: in <module>
from selenium import webdriver
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/__init__.py:18: in <module>
from .firefox.webdriver import WebDriver as Firefox # noqa
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/firefox/webdriver.py:26: in <module>
from selenium.webdriver.remote.webdriver import WebDriver as RemoteWebDriver
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py:52: in <module>
from .remote_connection import RemoteConnection
/opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/remote_connection.py:25: in <module>
import certifi
E ModuleNotFoundError: No module named 'certifi'
```
### Operating System
Ubuntu
### Selenium version
Python 4.4.1
### What are the browser(s) and version(s) where you see this issue?
None
### What are the browser driver(s) and version(s) where you see this issue?
None
### Are you using Selenium Grid?
_No response_ | non_infrastructure | modulenotfounderror no module named certifi what happened error in how can we reproduce the issue shell from selenium import webdriver relevant log output shell test py in from selenium import webdriver opt hostedtoolcache python lib site packages selenium webdriver init py in from firefox webdriver import webdriver as firefox noqa opt hostedtoolcache python lib site packages selenium webdriver firefox webdriver py in from selenium webdriver remote webdriver import webdriver as remotewebdriver opt hostedtoolcache python lib site packages selenium webdriver remote webdriver py in from remote connection import remoteconnection opt hostedtoolcache python lib site packages selenium webdriver remote remote connection py in import certifi e modulenotfounderror no module named certifi operating system ubuntu selenium version python what are the browser s and version s where you see this issue none what are the browser driver s and version s where you see this issue none are you using selenium grid no response | 0 |
3,988 | 4,754,951,420 | IssuesEvent | 2016-10-24 09:11:42 | robotology/QA | https://api.github.com/repos/robotology/QA | closed | New GUI ethLoader | firmware icub-ethernet infrastructure | Hi,
I wanted to update the robot, and as I've understood from issues #134 and #135, there is a new gui for the `ethLoader`, and all the updated instructions are referred to the new gui.
However, even if I updated everything, I still see the old gui; therefore, is there any option or so that I had to activate to see the new gui?
(I updated the firmware anyway with the old one, but for future and also students it would be better to have the one that matches with the instructions).
Thanks! | 1.0 | New GUI ethLoader - Hi,
I wanted to update the robot, and as I've understood from issues #134 and #135, there is a new gui for the `ethLoader`, and all the updated instructions are referred to the new gui.
However, even if I updated everything, I still see the old gui; therefore, is there any option or so that I had to activate to see the new gui?
(I updated the firmware anyway with the old one, but for future and also students it would be better to have the one that matches with the instructions).
Thanks! | infrastructure | new gui ethloader hi i wanted to update the robot and as i ve understood from issues and there is a new gui for the ethloader and all the updated instructions are referred to the new gui however even if i updated everything i still see the old gui therefore is there any option or so that i had to activate to see the new gui i updated the firmware anyway with the old one but for future and also students it would be better to have the one that matches with the instructions thanks | 1 |
9,060 | 3,834,080,298 | IssuesEvent | 2016-04-01 08:11:28 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Joomla 3.5 Admin Problem | No Code Attached Yet | Just upgraded to 3.5, now I can only see active position items under the "Select Position" drop down menu, all other positions are gone. See attached screen shot. HELP! | 1.0 | Joomla 3.5 Admin Problem - Just upgraded to 3.5, now I can only see active position items under the "Select Position" drop down menu, all other positions are gone. See attached screen shot. HELP! | non_infrastructure | joomla admin problem just upgraded to now i can only see active position items under the select position drop down menu all other positions are gone see attached screen shot help | 0 |
11,336 | 9,311,762,219 | IssuesEvent | 2019-03-25 22:21:56 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | Logged-in state is not consistent between Fetch data page and account management pages in React app with authentication in ASP.NET Core 3 Preview 3 | area-mvc identity-service investigate | ### Describe the bug
In a React app with authentication created using ASP.NET Core 3 Preview 3 when I click "Logout" on my account profile page I seem to be logged out there, but still logged in on the home page and "Fetch data" page.
### To Reproduce
1. Using ASP.NET Core 3.0 Preview 3 create a new project with `dotnet new react --auth Individual`
2. Apply the various workarounds at https://github.com/dotnet/core/blob/master/release-notes/3.0/preview/3.0.0-preview-known-issues.md
3. Run the app.
4. Click "Register" and complete the registration form (or log in with an existing account).
5. Click on "Hello youremail@..." in the header to go to /Identity/Account/Manage
6. Click "Logout" in the header on that page. You should be redirected back the homepage.
7. Observe that the header still says "Hello youremail@..." and shows a "Logout" link.
8. Click "Fetch data" in the header - the Weather forecast page is displayed, so you are indeed still logged in.
9. Click on "Hello youremail@..." in the header again. You are now prompted to log in, so you are not logged in on /Identity/Account/Manage.
### Expected behavior
When I click "Logout" on any page I am logged out for the entire application.
Note that the reverse is not true - when I log out from the home page and browse to /Identity/Account/Manage I am prompted to log in, as expected. | 1.0 | Logged-in state is not consistent between Fetch data page and account management pages in React app with authentication in ASP.NET Core 3 Preview 3 - ### Describe the bug
In a React app with authentication created using ASP.NET Core 3 Preview 3 when I click "Logout" on my account profile page I seem to be logged out there, but still logged in on the home page and "Fetch data" page.
### To Reproduce
1. Using ASP.NET Core 3.0 Preview 3 create a new project with `dotnet new react --auth Individual`
2. Apply the various workarounds at https://github.com/dotnet/core/blob/master/release-notes/3.0/preview/3.0.0-preview-known-issues.md
3. Run the app.
4. Click "Register" and complete the registration form (or log in with an existing account).
5. Click on "Hello youremail@..." in the header to go to /Identity/Account/Manage
6. Click "Logout" in the header on that page. You should be redirected back the homepage.
7. Observe that the header still says "Hello youremail@..." and shows a "Logout" link.
8. Click "Fetch data" in the header - the Weather forecast page is displayed, so you are indeed still logged in.
9. Click on "Hello youremail@..." in the header again. You are now prompted to log in, so you are not logged in on /Identity/Account/Manage.
### Expected behavior
When I click "Logout" on any page I am logged out for the entire application.
Note that the reverse is not true - when I log out from the home page and browse to /Identity/Account/Manage I am prompted to log in, as expected. | non_infrastructure | logged in state is not consistent between fetch data page and account management pages in react app with authentication in asp net core preview describe the bug in a react app with authentication created using asp net core preview when i click logout on my account profile page i seem to be logged out there but still logged in on the home page and fetch data page to reproduce using asp net core preview create a new project with dotnet new react auth individual apply the various workarounds at run the app click register and complete the registration form or log in with an existing account click on hello youremail in the header to go to identity account manage click logout in the header on that page you should be redirected back the homepage observe that the header still says hello youremail and shows a logout link click fetch data in the header the weather forecast page is displayed so you are indeed still logged in click on hello youremail in the header again you are now prompted to log in so you are not logged in on identity account manage expected behavior when i click logout on any page i am logged out for the entire application note that the reverse is not true when i log out from the home page and browse to identity account manage i am prompted to log in as expected | 0 |
22,954 | 15,696,571,661 | IssuesEvent | 2021-03-26 02:22:16 | seattle-uat/civiform | https://api.github.com/repos/seattle-uat/civiform | closed | enable civiform.com for our AWS instance | dev infrastructure | Can we point civiform.com to our AWS instance. I'm holding the domain under my personal GoDaddy account. Please provide me the configuration that I need to set.
Gerardo | 1.0 | enable civiform.com for our AWS instance - Can we point civiform.com to our AWS instance. I'm holding the domain under my personal GoDaddy account. Please provide me the configuration that I need to set.
Gerardo | infrastructure | enable civiform com for our aws instance can we point civiform com to our aws instance i m holding the domain under my personal godaddy account please provide me the configuration that i need to set gerardo | 1 |
28,925 | 23,613,870,699 | IssuesEvent | 2022-08-24 14:25:37 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | Include upper level classes plus bridging axioms in go-plus import chain | Infrastructure | Note: this is prompted by the desire to retire go-upper.obo https://github.com/geneontology/go-site/issues/129
Including upper level classes (primarily BFO but also CARO) in the import chain is itself straightforward. Note that we may want to limit confusing annotation assertions.
However, this is not much use unless bindings are present.
An early attempt for GO:
http://purl.obolibrary.org/obo/go/extensions/go-bfo-bridge.owl
this is probably overkill and can be replaced by 3 binding/bridge axioms (MF,BP->P, CC->CARO:?)
However, we will also need to ensure that bindings are present for all imported ontologies, to prevent them hanging out at root. There are a number of issues here:
- we might need to make ontological commitments on others behalf, e.g. SO
- import dependencies are hard to manage, see issue 38 in the [robot tracker](https://github.com/ontodev/robot/issues). but this is something we need to tackle anyway
This may wait til after:
#12211
| 1.0 | Include upper level classes plus bridging axioms in go-plus import chain - Note: this is prompted by the desire to retire go-upper.obo https://github.com/geneontology/go-site/issues/129
Including upper level classes (primarily BFO but also CARO) in the import chain is itself straightforward. Note that we may want to limit confusing annotation assertions.
However, this is not much use unless bindings are present.
An early attempt for GO:
http://purl.obolibrary.org/obo/go/extensions/go-bfo-bridge.owl
this is probably overkill and can be replaced by 3 binding/bridge axioms (MF,BP->P, CC->CARO:?)
However, we will also need to ensure that bindings are present for all imported ontologies, to prevent them hanging out at root. There are a number of issues here:
- we might need to make ontological commitments on others behalf, e.g. SO
- import dependencies are hard to manage, see issue 38 in the [robot tracker](https://github.com/ontodev/robot/issues). but this is something we need to tackle anyway
This may wait til after:
#12211
| infrastructure | include upper level classes plus bridging axioms in go plus import chain note this is prompted by the desire to retire go upper obo including upper level classes primarily bfo but also caro in the import chain is itself straightforward note that we may want to limit confusing annotation assertions however this is not much use unless bindings are present an early attempt for go this is probably overkill and can be replaced by binding bridge axioms mf bp p cc caro however we will also need to ensure that bindings are present for all imported ontologies to prevent them hanging out at root there are a number of issues here we might need to make ontological commitments on others behalf e g so import dependencies are hard to manage see issue in the but this is something we need to tackle anyway this may wait til after | 1 |
566,583 | 16,824,772,206 | IssuesEvent | 2021-06-17 17:01:25 | heyset/deskinar | https://api.github.com/repos/heyset/deskinar | opened | User should be able to answer questions | 3 points high priority new feature | Any user (for now) should be able to answer questions, and add them to the thread. | 1.0 | User should be able to answer questions - Any user (for now) should be able to answer questions, and add them to the thread. | non_infrastructure | user should be able to answer questions any user for now should be able to answer questions and add them to the thread | 0 |
13,911 | 10,543,972,778 | IssuesEvent | 2019-10-02 15:57:27 | mp-access/Backend | https://api.github.com/repos/mp-access/Backend | opened | User Acceptance Environment | Infrastructure | For our new release procedure we want a new user acceptance environment on digital ocean.
- We have a release candidate branch that is deployed on digital ocean.
- RC feature freeze is on Wednesday an is testes one week until the release one week later.
| 1.0 | User Acceptance Environment - For our new release procedure we want a new user acceptance environment on digital ocean.
- We have a release candidate branch that is deployed on digital ocean.
- RC feature freeze is on Wednesday an is testes one week until the release one week later.
| infrastructure | user acceptance environment for our new release procedure we want a new user acceptance environment on digital ocean we have a release candidate branch that is deployed on digital ocean rc feature freeze is on wednesday an is testes one week until the release one week later | 1 |
450,648 | 31,934,154,239 | IssuesEvent | 2023-09-19 09:22:07 | FuelLabs/sway-libs | https://api.github.com/repos/FuelLabs/sway-libs | closed | Specification missing function parameters | Documentation Good First Issue Improvement | ### Motivation
In all of the libraries i.e. [NFT](https://github.com/FuelLabs/sway-libs/blob/master/libs/nft/SPECIFICATION.md) are missing function parameters that tell the user how to use the specific function.
It has been brought up in the forum [here](https://forum.fuel.network/t/sway-libs-nft-library/2893/5) | 1.0 | Specification missing function parameters - ### Motivation
In all of the libraries i.e. [NFT](https://github.com/FuelLabs/sway-libs/blob/master/libs/nft/SPECIFICATION.md) are missing function parameters that tell the user how to use the specific function.
It has been brought up in the forum [here](https://forum.fuel.network/t/sway-libs-nft-library/2893/5) | non_infrastructure | specification missing function parameters motivation in all of the libraries i e are missing function parameters that tell the user how to use the specific function it has been brought up in the forum | 0 |
52,663 | 13,224,888,279 | IssuesEvent | 2020-08-17 20:03:04 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | i3_add_pybindings in top level CMakeLists.txt -> wrong lib location (Trac #120) | Migrated from Trac cmake defect | It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,
the build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib
A work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()
But Troy claims they should happily coexist at top level CMakeLists.txt
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/120">https://code.icecube.wisc.edu/projects/icecube/ticket/120</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"_ts": "1416713876900096",
"description": "It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,\nthe build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib\n\nA work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()\n\nBut Troy claims they should happily coexist at top level CMakeLists.txt",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2008-09-03T15:55:07",
"component": "cmake",
"summary": "i3_add_pybindings in top level CMakeLists.txt -> wrong lib location",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | i3_add_pybindings in top level CMakeLists.txt -> wrong lib location (Trac #120) - It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,
the build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib
A work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()
But Troy claims they should happily coexist at top level CMakeLists.txt
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/120">https://code.icecube.wisc.edu/projects/icecube/ticket/120</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"_ts": "1416713876900096",
"description": "It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,\nthe build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib\n\nA work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()\n\nBut Troy claims they should happily coexist at top level CMakeLists.txt",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2008-09-03T15:55:07",
"component": "cmake",
"summary": "i3_add_pybindings in top level CMakeLists.txt -> wrong lib location",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| non_infrastructure | add pybindings in top level cmakelists txt wrong lib location trac it seems when the add pybindings voodoo is in the toplevel cmakelists txt file the build paths of the standard libs is changed to build lib icecube instead of the standard build lib a work around is to use a separate cmakelists txt file in private pybindings and add the add subdirectory but troy claims they should happily coexist at top level cmakelists txt migrated from json status closed changetime ts description it seems when the add pybindings voodoo is in the toplevel cmakelists txt file nthe build paths of the standard libs is changed to build lib icecube instead of the standard build lib n na work around is to use a separate cmakelists txt file in private pybindings and add the add subdirectory n nbut troy claims they should happily coexist at top level cmakelists txt reporter blaufuss cc resolution fixed time component cmake summary add pybindings in top level cmakelists txt wrong lib location priority major keywords milestone owner troy type defect | 0 |
329,877 | 10,025,935,443 | IssuesEvent | 2019-07-17 04:33:34 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | none: Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use | cause/port-conflict co/none-driver kind/bug priority/backlog r/2019q2 triage/needs-information | <!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! -->
**The exact command to reproduce the issue**:
```
export MINIKUBE_WANTUPDATENOTIFICATION=false
export MINIKUBE_WANTREPORTERRORPROMPT=false
export CHANGE_MINIKUBE_NONE_USER=true
sudo -E minikube start --vm-driver=none
```
**The full output of the command that failed**:
```
😄 minikube v1.0.1 on linux (amd64)
🔥 Creating none VM (CPUs=2, Memory=2048MB, Disk=20000MB) ...
📶 "minikube" IP address is 192.168.8.159
🐳 Configuring Docker as the container runtime ...
🐳 Version of container runtime is 18.09.6
✨ Preparing Kubernetes environment ...
❌ Unable to load cached images: loading cached images: loading image /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: stat /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: no such file or directory
💾 Downloading kubeadm v1.14.1
💾 Downloading kubelet v1.14.1
🚜 Pulling images required by Kubernetes v1.14.1 ...
🚀 Launching Kubernetes v1.14.1 using kubeadm ...
💣 Error starting cluster: kubeadm init:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
: running command:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
output: [init] Using Kubernetes version: v1.14.1
[preflight] Running pre-flight checks
[WARNING FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml]: /etc/kubernetes/manifests/kube-apiserver.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml]: /etc/kubernetes/manifests/kube-controller-manager.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml]: /etc/kubernetes/manifests/kube-scheduler.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-etcd.yaml]: /etc/kubernetes/manifests/etcd.yaml already exists
[WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/
[WARNING Swap]: running with swap on is not supported. Please disable swap
[WARNING FileExisting-ebtables]: ebtables not found in system path
[WARNING FileExisting-ethtool]: ethtool not found in system path
[WARNING FileExisting-socat]: socat not found in system path
[WARNING Hostname]: hostname "minikube" could not be reached
[WARNING Hostname]: hostname "minikube": lookup minikube on 127.0.0.53:53: server misbehaving
[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
[WARNING Port-10250]: Port 10250 is in use
[WARNING DirAvailable--data-minikube]: /data/minikube is not empty
error execution phase preflight: [preflight] Some fatal errors occurred:
[ERROR Port-8443]: Port 8443 is in use
[ERROR Port-10251]: Port 10251 is in use
[ERROR Port-10252]: Port 10252 is in use
[ERROR Port-2379]: Port 2379 is in use
[ERROR Port-2380]: Port 2380 is in use
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
: running command:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
.: exit status 1
😿 Sorry that minikube crashed. If this was unexpected, we would love to hear from you:
👉 https://github.com/kubernetes/minikube/issues/new
❌ Problems detected in "kube-addon-manager":
error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused
```
**The output of the `minikube logs` command**:
```
==> dmesg <==
[May11 16:28] secureboot: Secure boot could not be determined (mode 0)
[ +0.000000] pmd_set_huge: Cannot satisfy [mem 0xf8000000-0xf8200000] with a huge-page mapping due to MTRR override.
[ +1.100563] r8169 0000:02:00.0: can't disable ASPM; OS doesn't have ASPM control
[ +0.333285] ata5.00: supports DRM functions and may not be fully accessible
[ +0.000140] ata5.00: READ LOG DMA EXT failed, trying PIO
[ +0.018473] ata5.00: supports DRM functions and may not be fully accessible
[ +0.650433] usb 3-5.2: device descriptor read/64, error -32
[ +9.859326] kauditd_printk_skb: 58 callbacks suppressed
[ +0.489483] aufs au_opts_verify:1609:dockerd[1937]: dirperm1 breaks the protection by the permission bits on the lower branch
[May12 03:01] IRQ 16: no longer affine to CPU1
[ +0.000007] IRQ 29: no longer affine to CPU1
[ +0.024017] IRQ 23: no longer affine to CPU2
[ +0.000007] IRQ 27: no longer affine to CPU2
[ +0.032058] IRQ 26: no longer affine to CPU3
[ +0.000010] IRQ 28: no longer affine to CPU3
[ +0.012294] cache: parent cpu1 should not be sleeping
[ +0.002224] cache: parent cpu2 should not be sleeping
[ +0.002101] cache: parent cpu3 should not be sleeping
[ +0.375470] ata5.00: supports DRM functions and may not be fully accessible
[ +0.018803] ata5.00: supports DRM functions and may not be fully accessible
[ +5.010804] ata1: link is slow to respond, please be patient (ready=0)
[ +4.651973] ata1: COMRESET failed (errno=-16)
[May12 06:52] sd 6:0:0:0: [sdc] No Caching mode page found
[ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through
[ +0.012436] sd 6:0:0:1: [sdd] No Caching mode page found
[ +0.000009] sd 6:0:0:1: [sdd] Assuming drive cache: write through
[May12 11:07] sd 6:0:0:0: [sdc] No Caching mode page found
[ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through
[ +0.001174] sd 6:0:0:1: [sdd] No Caching mode page found
[ +0.000003] sd 6:0:0:1: [sdd] Assuming drive cache: write through
[May12 12:03] kauditd_printk_skb: 32 callbacks suppressed
[May12 18:13] IRQ 16: no longer affine to CPU1
[ +0.000006] IRQ 29: no longer affine to CPU1
[ +0.024149] IRQ 23: no longer affine to CPU2
[ +0.000006] IRQ 27: no longer affine to CPU2
[ +0.032001] IRQ 26: no longer affine to CPU3
[ +0.000007] IRQ 28: no longer affine to CPU3
[ +0.011035] cache: parent cpu1 should not be sleeping
[ +0.002241] cache: parent cpu2 should not be sleeping
[ +0.002119] cache: parent cpu3 should not be sleeping
[ +0.369614] ata5.00: supports DRM functions and may not be fully accessible
[ +0.019113] ata5.00: supports DRM functions and may not be fully accessible
[ +5.037481] ata1: link is slow to respond, please be patient (ready=0)
[ +4.676005] ata1: COMRESET failed (errno=-16)
[May13 02:54] kauditd_printk_skb: 37 callbacks suppressed
==> kernel <==
20:38:38 up 2 days, 4:10, 1 user, load average: 0.96, 0.86, 0.82
Linux desktop 4.18.0-18-generic #19~18.04.1-Ubuntu SMP Fri Apr 5 10:22:13 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
==> kube-addon-manager <==
INFO: == Kubernetes addon manager started at 2019-05-14T00:33:24+00:00 with ADDON_CHECK_INTERVAL_SEC=60 ==
error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused
WRN: == Failed to start /opt/namespace.yaml in namespace at 2019-05-14T00:33:24+00:00. 99 tries remaining. ==
INFO: == Default service account in the kube-system namespace has token default-token-hr5j9 ==
find: '/etc/kubernetes/admission-controls': No such file or directory
INFO: == Entering periodical apply loop at 2019-05-14T00:33:29+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:33:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:33:31+00:00 ==
namespace/kube-system unchanged
INFO: == Successfully started /opt/namespace.yaml in namespace at 2019-05-14T00:33:34+00:00
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:34:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:34:30+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:35:30+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:35:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:36:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:36:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:37:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:37:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:38:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:38:30+00:00 ==
==> kube-apiserver <==
I0514 00:38:34.831507 1 log.go:172] http: TLS handshake error from 127.0.0.1:49094: remote error: tls: bad certificate
I0514 00:38:35.021488 1 log.go:172] http: TLS handshake error from 127.0.0.1:49096: remote error: tls: bad certificate
I0514 00:38:35.235536 1 log.go:172] http: TLS handshake error from 127.0.0.1:49100: remote error: tls: bad certificate
I0514 00:38:35.436229 1 log.go:172] http: TLS handshake error from 127.0.0.1:49102: remote error: tls: bad certificate
I0514 00:38:35.632126 1 log.go:172] http: TLS handshake error from 127.0.0.1:49104: remote error: tls: bad certificate
I0514 00:38:35.708998 1 log.go:172] http: TLS handshake error from 127.0.0.1:49106: remote error: tls: bad certificate
I0514 00:38:35.711809 1 log.go:172] http: TLS handshake error from 127.0.0.1:49108: remote error: tls: bad certificate
I0514 00:38:35.711850 1 log.go:172] http: TLS handshake error from 127.0.0.1:49110: remote error: tls: bad certificate
I0514 00:38:35.711876 1 log.go:172] http: TLS handshake error from 127.0.0.1:49112: remote error: tls: bad certificate
I0514 00:38:35.713858 1 log.go:172] http: TLS handshake error from 127.0.0.1:49114: remote error: tls: bad certificate
I0514 00:38:35.714111 1 log.go:172] http: TLS handshake error from 127.0.0.1:49116: remote error: tls: bad certificate
I0514 00:38:35.715785 1 log.go:172] http: TLS handshake error from 127.0.0.1:49118: remote error: tls: bad certificate
I0514 00:38:35.716377 1 log.go:172] http: TLS handshake error from 127.0.0.1:49124: remote error: tls: bad certificate
I0514 00:38:35.716620 1 log.go:172] http: TLS handshake error from 127.0.0.1:49120: remote error: tls: bad certificate
I0514 00:38:35.716870 1 log.go:172] http: TLS handshake error from 127.0.0.1:49122: remote error: tls: bad certificate
I0514 00:38:35.722733 1 log.go:172] http: TLS handshake error from 127.0.0.1:49126: remote error: tls: bad certificate
I0514 00:38:35.840634 1 log.go:172] http: TLS handshake error from 127.0.0.1:49128: remote error: tls: bad certificate
I0514 00:38:36.027863 1 log.go:172] http: TLS handshake error from 127.0.0.1:49130: remote error: tls: bad certificate
I0514 00:38:36.243197 1 log.go:172] http: TLS handshake error from 127.0.0.1:49132: remote error: tls: bad certificate
I0514 00:38:36.299517 1 log.go:172] http: TLS handshake error from 127.0.0.1:49134: remote error: tls: bad certificate
I0514 00:38:36.446919 1 log.go:172] http: TLS handshake error from 127.0.0.1:49136: remote error: tls: bad certificate
I0514 00:38:36.640301 1 log.go:172] http: TLS handshake error from 127.0.0.1:49138: remote error: tls: bad certificate
I0514 00:38:36.717558 1 log.go:172] http: TLS handshake error from 127.0.0.1:49140: remote error: tls: bad certificate
I0514 00:38:36.720112 1 log.go:172] http: TLS handshake error from 127.0.0.1:49150: remote error: tls: bad certificate
I0514 00:38:36.720144 1 log.go:172] http: TLS handshake error from 127.0.0.1:49144: remote error: tls: bad certificate
I0514 00:38:36.720168 1 log.go:172] http: TLS handshake error from 127.0.0.1:49142: remote error: tls: bad certificate
I0514 00:38:36.721405 1 log.go:172] http: TLS handshake error from 127.0.0.1:49146: remote error: tls: bad certificate
I0514 00:38:36.721835 1 log.go:172] http: TLS handshake error from 127.0.0.1:49152: remote error: tls: bad certificate
I0514 00:38:36.721908 1 log.go:172] http: TLS handshake error from 127.0.0.1:49148: remote error: tls: bad certificate
I0514 00:38:36.722280 1 log.go:172] http: TLS handshake error from 127.0.0.1:49154: remote error: tls: bad certificate
I0514 00:38:36.723077 1 log.go:172] http: TLS handshake error from 127.0.0.1:49156: remote error: tls: bad certificate
I0514 00:38:36.723116 1 log.go:172] http: TLS handshake error from 127.0.0.1:49158: remote error: tls: bad certificate
I0514 00:38:36.843667 1 log.go:172] http: TLS handshake error from 127.0.0.1:49160: remote error: tls: bad certificate
I0514 00:38:37.032639 1 log.go:172] http: TLS handshake error from 127.0.0.1:49164: remote error: tls: bad certificate
I0514 00:38:37.253082 1 log.go:172] http: TLS handshake error from 127.0.0.1:49166: remote error: tls: bad certificate
I0514 00:38:37.461185 1 log.go:172] http: TLS handshake error from 127.0.0.1:49168: remote error: tls: bad certificate
I0514 00:38:37.647529 1 log.go:172] http: TLS handshake error from 127.0.0.1:49170: remote error: tls: bad certificate
I0514 00:38:37.726883 1 log.go:172] http: TLS handshake error from 127.0.0.1:49172: remote error: tls: bad certificate
I0514 00:38:37.731172 1 log.go:172] http: TLS handshake error from 127.0.0.1:49178: remote error: tls: bad certificate
I0514 00:38:37.731797 1 log.go:172] http: TLS handshake error from 127.0.0.1:49174: remote error: tls: bad certificate
I0514 00:38:37.731833 1 log.go:172] http: TLS handshake error from 127.0.0.1:49184: remote error: tls: bad certificate
I0514 00:38:37.731857 1 log.go:172] http: TLS handshake error from 127.0.0.1:49182: remote error: tls: bad certificate
I0514 00:38:37.731881 1 log.go:172] http: TLS handshake error from 127.0.0.1:49176: remote error: tls: bad certificate
I0514 00:38:37.733898 1 log.go:172] http: TLS handshake error from 127.0.0.1:49186: remote error: tls: bad certificate
I0514 00:38:37.733938 1 log.go:172] http: TLS handshake error from 127.0.0.1:49180: remote error: tls: bad certificate
I0514 00:38:37.734391 1 log.go:172] http: TLS handshake error from 127.0.0.1:49190: remote error: tls: bad certificate
I0514 00:38:37.734678 1 log.go:172] http: TLS handshake error from 127.0.0.1:49188: remote error: tls: bad certificate
I0514 00:38:37.847115 1 log.go:172] http: TLS handshake error from 127.0.0.1:49194: remote error: tls: bad certificate
I0514 00:38:38.035683 1 log.go:172] http: TLS handshake error from 127.0.0.1:49198: remote error: tls: bad certificate
I0514 00:38:38.202140 1 log.go:172] http: TLS handshake error from 127.0.0.1:49200: remote error: tls: bad certificate
==> kube-scheduler <==
E0514 00:38:33.696199 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: Get https://localhost:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.696962 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: Get https://localhost:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.696988 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: Get https://localhost:8443/api/v1/pods?fieldSelector=status.phase%3DFailed%!C(MISSING)status.phase%3DSucceeded&limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.699704 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: Get https://localhost:8443/api/v1/services?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700091 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: Get https://localhost:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700353 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: Get https://localhost:8443/apis/apps/v1/statefulsets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700372 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: Get https://localhost:8443/apis/apps/v1/replicasets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700394 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: Get https://localhost:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.701981 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: Get https://localhost:8443/api/v1/nodes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
```
**The operating system version**:
`Linux Mint 19.1 Tessa`
| 1.0 | none: Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use - <!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! -->
**The exact command to reproduce the issue**:
```
export MINIKUBE_WANTUPDATENOTIFICATION=false
export MINIKUBE_WANTREPORTERRORPROMPT=false
export CHANGE_MINIKUBE_NONE_USER=true
sudo -E minikube start --vm-driver=none
```
**The full output of the command that failed**:
```
😄 minikube v1.0.1 on linux (amd64)
🔥 Creating none VM (CPUs=2, Memory=2048MB, Disk=20000MB) ...
📶 "minikube" IP address is 192.168.8.159
🐳 Configuring Docker as the container runtime ...
🐳 Version of container runtime is 18.09.6
✨ Preparing Kubernetes environment ...
❌ Unable to load cached images: loading cached images: loading image /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: stat /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: no such file or directory
💾 Downloading kubeadm v1.14.1
💾 Downloading kubelet v1.14.1
🚜 Pulling images required by Kubernetes v1.14.1 ...
🚀 Launching Kubernetes v1.14.1 using kubeadm ...
💣 Error starting cluster: kubeadm init:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
: running command:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
output: [init] Using Kubernetes version: v1.14.1
[preflight] Running pre-flight checks
[WARNING FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml]: /etc/kubernetes/manifests/kube-apiserver.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml]: /etc/kubernetes/manifests/kube-controller-manager.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml]: /etc/kubernetes/manifests/kube-scheduler.yaml already exists
[WARNING FileAvailable--etc-kubernetes-manifests-etcd.yaml]: /etc/kubernetes/manifests/etcd.yaml already exists
[WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/
[WARNING Swap]: running with swap on is not supported. Please disable swap
[WARNING FileExisting-ebtables]: ebtables not found in system path
[WARNING FileExisting-ethtool]: ethtool not found in system path
[WARNING FileExisting-socat]: socat not found in system path
[WARNING Hostname]: hostname "minikube" could not be reached
[WARNING Hostname]: hostname "minikube": lookup minikube on 127.0.0.53:53: server misbehaving
[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
[WARNING Port-10250]: Port 10250 is in use
[WARNING DirAvailable--data-minikube]: /data/minikube is not empty
error execution phase preflight: [preflight] Some fatal errors occurred:
[ERROR Port-8443]: Port 8443 is in use
[ERROR Port-10251]: Port 10251 is in use
[ERROR Port-10252]: Port 10252 is in use
[ERROR Port-2379]: Port 2379 is in use
[ERROR Port-2380]: Port 2380 is in use
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
: running command:
sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI
.: exit status 1
😿 Sorry that minikube crashed. If this was unexpected, we would love to hear from you:
👉 https://github.com/kubernetes/minikube/issues/new
❌ Problems detected in "kube-addon-manager":
error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused
```
**The output of the `minikube logs` command**:
```
==> dmesg <==
[May11 16:28] secureboot: Secure boot could not be determined (mode 0)
[ +0.000000] pmd_set_huge: Cannot satisfy [mem 0xf8000000-0xf8200000] with a huge-page mapping due to MTRR override.
[ +1.100563] r8169 0000:02:00.0: can't disable ASPM; OS doesn't have ASPM control
[ +0.333285] ata5.00: supports DRM functions and may not be fully accessible
[ +0.000140] ata5.00: READ LOG DMA EXT failed, trying PIO
[ +0.018473] ata5.00: supports DRM functions and may not be fully accessible
[ +0.650433] usb 3-5.2: device descriptor read/64, error -32
[ +9.859326] kauditd_printk_skb: 58 callbacks suppressed
[ +0.489483] aufs au_opts_verify:1609:dockerd[1937]: dirperm1 breaks the protection by the permission bits on the lower branch
[May12 03:01] IRQ 16: no longer affine to CPU1
[ +0.000007] IRQ 29: no longer affine to CPU1
[ +0.024017] IRQ 23: no longer affine to CPU2
[ +0.000007] IRQ 27: no longer affine to CPU2
[ +0.032058] IRQ 26: no longer affine to CPU3
[ +0.000010] IRQ 28: no longer affine to CPU3
[ +0.012294] cache: parent cpu1 should not be sleeping
[ +0.002224] cache: parent cpu2 should not be sleeping
[ +0.002101] cache: parent cpu3 should not be sleeping
[ +0.375470] ata5.00: supports DRM functions and may not be fully accessible
[ +0.018803] ata5.00: supports DRM functions and may not be fully accessible
[ +5.010804] ata1: link is slow to respond, please be patient (ready=0)
[ +4.651973] ata1: COMRESET failed (errno=-16)
[May12 06:52] sd 6:0:0:0: [sdc] No Caching mode page found
[ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through
[ +0.012436] sd 6:0:0:1: [sdd] No Caching mode page found
[ +0.000009] sd 6:0:0:1: [sdd] Assuming drive cache: write through
[May12 11:07] sd 6:0:0:0: [sdc] No Caching mode page found
[ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through
[ +0.001174] sd 6:0:0:1: [sdd] No Caching mode page found
[ +0.000003] sd 6:0:0:1: [sdd] Assuming drive cache: write through
[May12 12:03] kauditd_printk_skb: 32 callbacks suppressed
[May12 18:13] IRQ 16: no longer affine to CPU1
[ +0.000006] IRQ 29: no longer affine to CPU1
[ +0.024149] IRQ 23: no longer affine to CPU2
[ +0.000006] IRQ 27: no longer affine to CPU2
[ +0.032001] IRQ 26: no longer affine to CPU3
[ +0.000007] IRQ 28: no longer affine to CPU3
[ +0.011035] cache: parent cpu1 should not be sleeping
[ +0.002241] cache: parent cpu2 should not be sleeping
[ +0.002119] cache: parent cpu3 should not be sleeping
[ +0.369614] ata5.00: supports DRM functions and may not be fully accessible
[ +0.019113] ata5.00: supports DRM functions and may not be fully accessible
[ +5.037481] ata1: link is slow to respond, please be patient (ready=0)
[ +4.676005] ata1: COMRESET failed (errno=-16)
[May13 02:54] kauditd_printk_skb: 37 callbacks suppressed
==> kernel <==
20:38:38 up 2 days, 4:10, 1 user, load average: 0.96, 0.86, 0.82
Linux desktop 4.18.0-18-generic #19~18.04.1-Ubuntu SMP Fri Apr 5 10:22:13 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
==> kube-addon-manager <==
INFO: == Kubernetes addon manager started at 2019-05-14T00:33:24+00:00 with ADDON_CHECK_INTERVAL_SEC=60 ==
error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused
WRN: == Failed to start /opt/namespace.yaml in namespace at 2019-05-14T00:33:24+00:00. 99 tries remaining. ==
INFO: == Default service account in the kube-system namespace has token default-token-hr5j9 ==
find: '/etc/kubernetes/admission-controls': No such file or directory
INFO: == Entering periodical apply loop at 2019-05-14T00:33:29+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:33:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:33:31+00:00 ==
namespace/kube-system unchanged
INFO: == Successfully started /opt/namespace.yaml in namespace at 2019-05-14T00:33:34+00:00
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:34:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:34:30+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:35:30+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:35:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:36:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:36:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:37:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:37:31+00:00 ==
INFO: Leader is desktop
INFO: == Kubernetes addon ensure completed at 2019-05-14T00:38:29+00:00 ==
INFO: == Reconciling with deprecated label ==
error: no objects passed to apply
INFO: == Reconciling with addon-manager label ==
serviceaccount/storage-provisioner unchanged
INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:38:30+00:00 ==
==> kube-apiserver <==
I0514 00:38:34.831507 1 log.go:172] http: TLS handshake error from 127.0.0.1:49094: remote error: tls: bad certificate
I0514 00:38:35.021488 1 log.go:172] http: TLS handshake error from 127.0.0.1:49096: remote error: tls: bad certificate
I0514 00:38:35.235536 1 log.go:172] http: TLS handshake error from 127.0.0.1:49100: remote error: tls: bad certificate
I0514 00:38:35.436229 1 log.go:172] http: TLS handshake error from 127.0.0.1:49102: remote error: tls: bad certificate
I0514 00:38:35.632126 1 log.go:172] http: TLS handshake error from 127.0.0.1:49104: remote error: tls: bad certificate
I0514 00:38:35.708998 1 log.go:172] http: TLS handshake error from 127.0.0.1:49106: remote error: tls: bad certificate
I0514 00:38:35.711809 1 log.go:172] http: TLS handshake error from 127.0.0.1:49108: remote error: tls: bad certificate
I0514 00:38:35.711850 1 log.go:172] http: TLS handshake error from 127.0.0.1:49110: remote error: tls: bad certificate
I0514 00:38:35.711876 1 log.go:172] http: TLS handshake error from 127.0.0.1:49112: remote error: tls: bad certificate
I0514 00:38:35.713858 1 log.go:172] http: TLS handshake error from 127.0.0.1:49114: remote error: tls: bad certificate
I0514 00:38:35.714111 1 log.go:172] http: TLS handshake error from 127.0.0.1:49116: remote error: tls: bad certificate
I0514 00:38:35.715785 1 log.go:172] http: TLS handshake error from 127.0.0.1:49118: remote error: tls: bad certificate
I0514 00:38:35.716377 1 log.go:172] http: TLS handshake error from 127.0.0.1:49124: remote error: tls: bad certificate
I0514 00:38:35.716620 1 log.go:172] http: TLS handshake error from 127.0.0.1:49120: remote error: tls: bad certificate
I0514 00:38:35.716870 1 log.go:172] http: TLS handshake error from 127.0.0.1:49122: remote error: tls: bad certificate
I0514 00:38:35.722733 1 log.go:172] http: TLS handshake error from 127.0.0.1:49126: remote error: tls: bad certificate
I0514 00:38:35.840634 1 log.go:172] http: TLS handshake error from 127.0.0.1:49128: remote error: tls: bad certificate
I0514 00:38:36.027863 1 log.go:172] http: TLS handshake error from 127.0.0.1:49130: remote error: tls: bad certificate
I0514 00:38:36.243197 1 log.go:172] http: TLS handshake error from 127.0.0.1:49132: remote error: tls: bad certificate
I0514 00:38:36.299517 1 log.go:172] http: TLS handshake error from 127.0.0.1:49134: remote error: tls: bad certificate
I0514 00:38:36.446919 1 log.go:172] http: TLS handshake error from 127.0.0.1:49136: remote error: tls: bad certificate
I0514 00:38:36.640301 1 log.go:172] http: TLS handshake error from 127.0.0.1:49138: remote error: tls: bad certificate
I0514 00:38:36.717558 1 log.go:172] http: TLS handshake error from 127.0.0.1:49140: remote error: tls: bad certificate
I0514 00:38:36.720112 1 log.go:172] http: TLS handshake error from 127.0.0.1:49150: remote error: tls: bad certificate
I0514 00:38:36.720144 1 log.go:172] http: TLS handshake error from 127.0.0.1:49144: remote error: tls: bad certificate
I0514 00:38:36.720168 1 log.go:172] http: TLS handshake error from 127.0.0.1:49142: remote error: tls: bad certificate
I0514 00:38:36.721405 1 log.go:172] http: TLS handshake error from 127.0.0.1:49146: remote error: tls: bad certificate
I0514 00:38:36.721835 1 log.go:172] http: TLS handshake error from 127.0.0.1:49152: remote error: tls: bad certificate
I0514 00:38:36.721908 1 log.go:172] http: TLS handshake error from 127.0.0.1:49148: remote error: tls: bad certificate
I0514 00:38:36.722280 1 log.go:172] http: TLS handshake error from 127.0.0.1:49154: remote error: tls: bad certificate
I0514 00:38:36.723077 1 log.go:172] http: TLS handshake error from 127.0.0.1:49156: remote error: tls: bad certificate
I0514 00:38:36.723116 1 log.go:172] http: TLS handshake error from 127.0.0.1:49158: remote error: tls: bad certificate
I0514 00:38:36.843667 1 log.go:172] http: TLS handshake error from 127.0.0.1:49160: remote error: tls: bad certificate
I0514 00:38:37.032639 1 log.go:172] http: TLS handshake error from 127.0.0.1:49164: remote error: tls: bad certificate
I0514 00:38:37.253082 1 log.go:172] http: TLS handshake error from 127.0.0.1:49166: remote error: tls: bad certificate
I0514 00:38:37.461185 1 log.go:172] http: TLS handshake error from 127.0.0.1:49168: remote error: tls: bad certificate
I0514 00:38:37.647529 1 log.go:172] http: TLS handshake error from 127.0.0.1:49170: remote error: tls: bad certificate
I0514 00:38:37.726883 1 log.go:172] http: TLS handshake error from 127.0.0.1:49172: remote error: tls: bad certificate
I0514 00:38:37.731172 1 log.go:172] http: TLS handshake error from 127.0.0.1:49178: remote error: tls: bad certificate
I0514 00:38:37.731797 1 log.go:172] http: TLS handshake error from 127.0.0.1:49174: remote error: tls: bad certificate
I0514 00:38:37.731833 1 log.go:172] http: TLS handshake error from 127.0.0.1:49184: remote error: tls: bad certificate
I0514 00:38:37.731857 1 log.go:172] http: TLS handshake error from 127.0.0.1:49182: remote error: tls: bad certificate
I0514 00:38:37.731881 1 log.go:172] http: TLS handshake error from 127.0.0.1:49176: remote error: tls: bad certificate
I0514 00:38:37.733898 1 log.go:172] http: TLS handshake error from 127.0.0.1:49186: remote error: tls: bad certificate
I0514 00:38:37.733938 1 log.go:172] http: TLS handshake error from 127.0.0.1:49180: remote error: tls: bad certificate
I0514 00:38:37.734391 1 log.go:172] http: TLS handshake error from 127.0.0.1:49190: remote error: tls: bad certificate
I0514 00:38:37.734678 1 log.go:172] http: TLS handshake error from 127.0.0.1:49188: remote error: tls: bad certificate
I0514 00:38:37.847115 1 log.go:172] http: TLS handshake error from 127.0.0.1:49194: remote error: tls: bad certificate
I0514 00:38:38.035683 1 log.go:172] http: TLS handshake error from 127.0.0.1:49198: remote error: tls: bad certificate
I0514 00:38:38.202140 1 log.go:172] http: TLS handshake error from 127.0.0.1:49200: remote error: tls: bad certificate
==> kube-scheduler <==
E0514 00:38:33.696199 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: Get https://localhost:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.696962 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: Get https://localhost:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.696988 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: Get https://localhost:8443/api/v1/pods?fieldSelector=status.phase%3DFailed%!C(MISSING)status.phase%3DSucceeded&limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.699704 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: Get https://localhost:8443/api/v1/services?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700091 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: Get https://localhost:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700353 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: Get https://localhost:8443/apis/apps/v1/statefulsets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700372 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: Get https://localhost:8443/apis/apps/v1/replicasets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.700394 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: Get https://localhost:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
E0514 00:38:33.701981 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: Get https://localhost:8443/api/v1/nodes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA")
```
**The operating system version**:
`Linux Mint 19.1 Tessa`
| non_infrastructure | none some fatal errors occurred port is in use the exact command to reproduce the issue export minikube wantupdatenotification false export minikube wantreporterrorprompt false export change minikube none user true sudo e minikube start vm driver none the full output of the command that failed 😄 minikube on linux 🔥 creating none vm cpus memory disk 📶 minikube ip address is 🐳 configuring docker as the container runtime 🐳 version of container runtime is ✨ preparing kubernetes environment ❌ unable to load cached images loading cached images loading image home vadim minikube cache images gcr io kube proxy stat home vadim minikube cache images gcr io kube proxy no such file or directory 💾 downloading kubeadm 💾 downloading kubelet 🚜 pulling images required by kubernetes 🚀 launching kubernetes using kubeadm 💣 error starting cluster kubeadm init sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri running command sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri output using kubernetes version running pre flight checks etc kubernetes manifests kube apiserver yaml already exists etc kubernetes manifests kube controller manager yaml already exists etc kubernetes manifests kube scheduler yaml already exists etc kubernetes manifests etcd yaml already exists detected cgroupfs as the docker cgroup driver the recommended driver is systemd please follow the guide at running with swap on is not supported please disable swap ebtables not found in system path ethtool not found in system path socat not found in system path hostname minikube could not be reached hostname minikube lookup minikube on server misbehaving kubelet service is not enabled please run systemctl enable kubelet service port is in use data minikube is not empty error execution phase preflight some fatal errors occurred port is in use port is in use port is in use port is in use port is in use if you know what you are doing you can make a check non fatal with ignore preflight errors running command sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri exit status 😿 sorry that minikube crashed if this was unexpected we would love to hear from you 👉 ❌ problems detected in kube addon manager error unable to recognize stdin get dial tcp connect connection refused the output of the minikube logs command dmesg secureboot secure boot could not be determined mode pmd set huge cannot satisfy with a huge page mapping due to mtrr override can t disable aspm os doesn t have aspm control supports drm functions and may not be fully accessible read log dma ext failed trying pio supports drm functions and may not be fully accessible usb device descriptor read error kauditd printk skb callbacks suppressed aufs au opts verify dockerd breaks the protection by the permission bits on the lower branch irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to cache parent should not be sleeping cache parent should not be sleeping cache parent should not be sleeping supports drm functions and may not be fully accessible supports drm functions and may not be fully accessible link is slow to respond please be patient ready comreset failed errno sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through kauditd printk skb callbacks suppressed irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to cache parent should not be sleeping cache parent should not be sleeping cache parent should not be sleeping supports drm functions and may not be fully accessible supports drm functions and may not be fully accessible link is slow to respond please be patient ready comreset failed errno kauditd printk skb callbacks suppressed kernel up days user load average linux desktop generic ubuntu smp fri apr utc gnu linux kube addon manager info kubernetes addon manager started at with addon check interval sec error unable to recognize stdin get dial tcp connect connection refused wrn failed to start opt namespace yaml in namespace at tries remaining info default service account in the kube system namespace has token default token find etc kubernetes admission controls no such file or directory info entering periodical apply loop at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at namespace kube system unchanged info successfully started opt namespace yaml in namespace at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at kube apiserver log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate kube scheduler reflector go io client go informers factory go failed to list persistentvolumeclaim get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list persistentvolume get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io kubernetes cmd kube scheduler app server go failed to list pod get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list service get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list poddisruptionbudget get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list statefulset get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list replicaset get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list storageclass get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list node get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca the operating system version linux mint tessa | 0 |
24,046 | 7,451,634,520 | IssuesEvent | 2018-03-29 04:12:35 | cerner/terra-core | https://api.github.com/repos/cerner/terra-core | closed | Update Express Dependency to ^4.16.0 | build-tools | # Issue Description
When visiting the `express` doc site, there is a notice stating:
> Express 4.16.0 contains important security updates.
For more information on what was added in this release, see the [4.16.0 changelog](http://expressjs.com/en/changelog/4x.html#4.16.0).
## Issue Type
<!-- Is this a new feature request, enhancement, bug report, other? -->
- [ ] New Feature
- [ ] Enhancement
- [ ] Bug
- [x] Other
| 1.0 | Update Express Dependency to ^4.16.0 - # Issue Description
When visiting the `express` doc site, there is a notice stating:
> Express 4.16.0 contains important security updates.
For more information on what was added in this release, see the [4.16.0 changelog](http://expressjs.com/en/changelog/4x.html#4.16.0).
## Issue Type
<!-- Is this a new feature request, enhancement, bug report, other? -->
- [ ] New Feature
- [ ] Enhancement
- [ ] Bug
- [x] Other
| non_infrastructure | update express dependency to issue description when visiting the express doc site there is a notice stating express contains important security updates for more information on what was added in this release see the issue type new feature enhancement bug other | 0 |
9,615 | 8,053,612,178 | IssuesEvent | 2018-08-02 00:01:47 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | opened | Change the different types of links to be a single [Link] attribute | interface/infrastructure refactor | Currently we have [Link], [LinkByPath], [ScopedLinkByName], [ScopedLink], [ChildLink], [ChildLinkByName], [ParentLink]. I think we need a single [Link] class with perhaps an enum argument e.g.
[Link(type=LinkType.Path, path="xyz")]
[Link(type=LinkType.Scoped, name="xyz")]
[Link(type=LinkType.Child, name="xyz")]
[Link(type=LinkType.Parent, name="xyz")] | 1.0 | Change the different types of links to be a single [Link] attribute - Currently we have [Link], [LinkByPath], [ScopedLinkByName], [ScopedLink], [ChildLink], [ChildLinkByName], [ParentLink]. I think we need a single [Link] class with perhaps an enum argument e.g.
[Link(type=LinkType.Path, path="xyz")]
[Link(type=LinkType.Scoped, name="xyz")]
[Link(type=LinkType.Child, name="xyz")]
[Link(type=LinkType.Parent, name="xyz")] | infrastructure | change the different types of links to be a single attribute currently we have i think we need a single class with perhaps an enum argument e g | 1 |
1,690 | 3,343,384,822 | IssuesEvent | 2015-11-15 13:11:44 | rackt/redux | https://api.github.com/repos/rackt/redux | closed | [React Native] Upgrading to Babel 6 and/or adding .babelrc to .npmignore | help wanted infrastructure | React Native will be shipping with Babel 6 in about three weeks. One consequence of this is that Babel finds .babelrc in `node_modules/redux` and tries to consume it -- since Redux is using Babel 5's configuration, this causes Babel 6 to crash. The current verdict is that packages will have to exclude .babelrc when distributed through npm: https://github.com/facebook/react-native/issues/4062. | 1.0 | [React Native] Upgrading to Babel 6 and/or adding .babelrc to .npmignore - React Native will be shipping with Babel 6 in about three weeks. One consequence of this is that Babel finds .babelrc in `node_modules/redux` and tries to consume it -- since Redux is using Babel 5's configuration, this causes Babel 6 to crash. The current verdict is that packages will have to exclude .babelrc when distributed through npm: https://github.com/facebook/react-native/issues/4062. | infrastructure | upgrading to babel and or adding babelrc to npmignore react native will be shipping with babel in about three weeks one consequence of this is that babel finds babelrc in node modules redux and tries to consume it since redux is using babel s configuration this causes babel to crash the current verdict is that packages will have to exclude babelrc when distributed through npm | 1 |
32,661 | 26,877,842,432 | IssuesEvent | 2023-02-05 08:56:18 | aimclub/FEDOT | https://api.github.com/repos/aimclub/FEDOT | closed | Make experiment for calculating overhead in using multiprocessing | research infrastructure | Perform experiment:
1) Run fitting n pipelines in multiprocessing mode
2) Run fitting n pipelines in single process mode
3) Calculate overhead for creating a process
| 1.0 | Make experiment for calculating overhead in using multiprocessing - Perform experiment:
1) Run fitting n pipelines in multiprocessing mode
2) Run fitting n pipelines in single process mode
3) Calculate overhead for creating a process
| infrastructure | make experiment for calculating overhead in using multiprocessing perform experiment run fitting n pipelines in multiprocessing mode run fitting n pipelines in single process mode calculate overhead for creating a process | 1 |
30,901 | 25,153,802,446 | IssuesEvent | 2022-11-10 12:03:30 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Improve concurrency of `runtest.sh` | os-linux test-enhancement area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate | Our unix test script `runtest.sh` runs TCs in parallel. However it looks a little inefficient for some scenarios.
For example, the script decides to run 4 processes concurrently. Then it starts first 4 tests. Let's number them 0 to 3. It waits for process 0. So if process 0 is long running but 1, 2 and 3 are not, process 0 could be a bottleneck. Even if 1, 2 and 3 are finished the script cannot run more processes before 0 is done. Because it waits for process sequentially.
I think it would improve concurrency of unit tests if it can wait for any of child processes, not a in sequential order. | 1.0 | Improve concurrency of `runtest.sh` - Our unix test script `runtest.sh` runs TCs in parallel. However it looks a little inefficient for some scenarios.
For example, the script decides to run 4 processes concurrently. Then it starts first 4 tests. Let's number them 0 to 3. It waits for process 0. So if process 0 is long running but 1, 2 and 3 are not, process 0 could be a bottleneck. Even if 1, 2 and 3 are finished the script cannot run more processes before 0 is done. Because it waits for process sequentially.
I think it would improve concurrency of unit tests if it can wait for any of child processes, not a in sequential order. | infrastructure | improve concurrency of runtest sh our unix test script runtest sh runs tcs in parallel however it looks a little inefficient for some scenarios for example the script decides to run processes concurrently then it starts first tests let s number them to it waits for process so if process is long running but and are not process could be a bottleneck even if and are finished the script cannot run more processes before is done because it waits for process sequentially i think it would improve concurrency of unit tests if it can wait for any of child processes not a in sequential order | 1 |
147,124 | 19,500,433,672 | IssuesEvent | 2021-12-28 01:31:33 | talevy013/TestTal | https://api.github.com/repos/talevy013/TestTal | opened | CVE-2020-10968 (High) detected in jackson-databind-2.6.7.1.jar | security vulnerability | ## CVE-2020-10968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.7.1/jackson-databind-2.6.7.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/talevy013/TestTal/commit/f001f7f069d9289dded859e738eb111a8fd2e984">f001f7f069d9289dded859e738eb111a8fd2e984</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy).
<p>Publish Date: 2020-03-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p>
<p>Release Date: 2020-03-26</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
| True | CVE-2020-10968 (High) detected in jackson-databind-2.6.7.1.jar - ## CVE-2020-10968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.7.1/jackson-databind-2.6.7.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/talevy013/TestTal/commit/f001f7f069d9289dded859e738eb111a8fd2e984">f001f7f069d9289dded859e738eb111a8fd2e984</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy).
<p>Publish Date: 2020-03-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p>
<p>Release Date: 2020-03-26</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
| non_infrastructure | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org aoju bus proxy provider remoting rmiprovider aka bus proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind | 0 |
1,382 | 3,179,709,491 | IssuesEvent | 2015-09-25 04:03:39 | t3kt/vjzual2 | https://api.github.com/repos/t3kt/vjzual2 | opened | support modules disabling header buttons | enhancement infrastructure ui | like not showing the toggle for advanced parameters (or viewers) when the module doesn't have any | 1.0 | support modules disabling header buttons - like not showing the toggle for advanced parameters (or viewers) when the module doesn't have any | infrastructure | support modules disabling header buttons like not showing the toggle for advanced parameters or viewers when the module doesn t have any | 1 |
27,765 | 30,338,864,610 | IssuesEvent | 2023-07-11 11:24:41 | ICTU/quality-time | https://api.github.com/repos/ICTU/quality-time | closed | Notify the user why editing is not possible | Usability | It would be useful to notify the user when the user tries to edit a field when it's not editable.
This can occur in the following situations:
- the user is not logged in
- the user is using the 'time machine' functionality where he/she looks into a view of the past, fields are not editable on purpose | True | Notify the user why editing is not possible - It would be useful to notify the user when the user tries to edit a field when it's not editable.
This can occur in the following situations:
- the user is not logged in
- the user is using the 'time machine' functionality where he/she looks into a view of the past, fields are not editable on purpose | non_infrastructure | notify the user why editing is not possible it would be useful to notify the user when the user tries to edit a field when it s not editable this can occur in the following situations the user is not logged in the user is using the time machine functionality where he she looks into a view of the past fields are not editable on purpose | 0 |
244,001 | 20,601,998,836 | IssuesEvent | 2022-03-06 12:10:36 | einTosti/BuildSystem | https://api.github.com/repos/einTosti/BuildSystem | closed | `NullPointerException` while opening GameRule inventory | type: bug status: needs testing version: 1.18 priority: high | ```
[12:50:35 ERROR]: Could not pass event InventoryClickEvent to BuildSystem v2.18
java.lang.NullPointerException: Cannot invoke "com.eintosti.buildsystem.version.GameRules.getInventory(org.bukkit.entity.Player, org.bukkit.World)" because "this.gameRules" is null
at com.eintosti.buildsystem.inventory.GameRuleInventory.openInventory(GameRuleInventory.java:45) ~[BuildSystem-2.18.jar:?]
at com.eintosti.buildsystem.inventory.EditInventory.onInventoryClick(EditInventory.java:285) ~[BuildSystem-2.18.jar:?]
at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor72.execute(Unknown Source) ~[?:?]
at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:75) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:git-Paper-215]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:629) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleContainerClick(ServerGamePacketListenerImpl.java:2838) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:58) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:13) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:56) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1423) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.c(MinecraftServer.java:189) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1400) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1393) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1371) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1277) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:317) ~[paper-1.18.1.jar:git-Paper-215]
at java.lang.Thread.run(Thread.java:833) ~[?:?]
``` | 1.0 | `NullPointerException` while opening GameRule inventory - ```
[12:50:35 ERROR]: Could not pass event InventoryClickEvent to BuildSystem v2.18
java.lang.NullPointerException: Cannot invoke "com.eintosti.buildsystem.version.GameRules.getInventory(org.bukkit.entity.Player, org.bukkit.World)" because "this.gameRules" is null
at com.eintosti.buildsystem.inventory.GameRuleInventory.openInventory(GameRuleInventory.java:45) ~[BuildSystem-2.18.jar:?]
at com.eintosti.buildsystem.inventory.EditInventory.onInventoryClick(EditInventory.java:285) ~[BuildSystem-2.18.jar:?]
at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor72.execute(Unknown Source) ~[?:?]
at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:75) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:git-Paper-215]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:629) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleContainerClick(ServerGamePacketListenerImpl.java:2838) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:58) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:13) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:56) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1423) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.c(MinecraftServer.java:189) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1400) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1393) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1371) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1277) ~[paper-1.18.1.jar:git-Paper-215]
at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:317) ~[paper-1.18.1.jar:git-Paper-215]
at java.lang.Thread.run(Thread.java:833) ~[?:?]
``` | non_infrastructure | nullpointerexception while opening gamerule inventory could not pass event inventoryclickevent to buildsystem java lang nullpointerexception cannot invoke com eintosti buildsystem version gamerules getinventory org bukkit entity player org bukkit world because this gamerules is null at com eintosti buildsystem inventory gameruleinventory openinventory gameruleinventory java at com eintosti buildsystem inventory editinventory oninventoryclick editinventory java at com destroystokyo paper event executor asm generated execute unknown source at org bukkit plugin eventexecutor lambda create eventexecutor java at co aikar timings timedeventexecutor execute timedeventexecutor java at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server network servergamepacketlistenerimpl handlecontainerclick servergamepacketlistenerimpl java at net minecraft network protocol game serverboundcontainerclickpacket handle serverboundcontainerclickpacket java at net minecraft network protocol game serverboundcontainerclickpacket handle serverboundcontainerclickpacket java at net minecraft network protocol packetutils lambda ensurerunningonsamethread packetutils java at net minecraft server ticktask run ticktask java at net minecraft util thread blockableeventloop doruntask blockableeventloop java at net minecraft util thread reentrantblockableeventloop doruntask reentrantblockableeventloop java at net minecraft server minecraftserver doruntask minecraftserver java at net minecraft server minecraftserver c minecraftserver java at net minecraft util thread blockableeventloop polltask blockableeventloop java at net minecraft server minecraftserver polltaskinternal minecraftserver java at net minecraft server minecraftserver polltask minecraftserver java at net minecraft util thread blockableeventloop managedblock blockableeventloop java at net minecraft server minecraftserver waituntilnexttick minecraftserver java at net minecraft server minecraftserver runserver minecraftserver java at net minecraft server minecraftserver lambda spin minecraftserver java at java lang thread run thread java | 0 |
130,151 | 18,042,014,278 | IssuesEvent | 2021-09-18 07:38:39 | m93a/filtrex | https://api.github.com/repos/m93a/filtrex | closed | Make errors i18n-friendly | enhancement design implemented | All errors, both thrown and compile time and returned at runtime, should have an identifier which would make translation possible. It is not acceptable to show English errors in a non-English application. | 1.0 | Make errors i18n-friendly - All errors, both thrown and compile time and returned at runtime, should have an identifier which would make translation possible. It is not acceptable to show English errors in a non-English application. | non_infrastructure | make errors friendly all errors both thrown and compile time and returned at runtime should have an identifier which would make translation possible it is not acceptable to show english errors in a non english application | 0 |
4,198 | 4,890,105,849 | IssuesEvent | 2016-11-18 12:41:39 | SatelliteQE/robottelo | https://api.github.com/repos/SatelliteQE/robottelo | closed | Workaround upstream bug closure in downstream jobs | Infrastructure | Skip downstream tests if
`bug status = 'verified' && QA whiteboard contains 'Verified in upstream'`.
| 1.0 | Workaround upstream bug closure in downstream jobs - Skip downstream tests if
`bug status = 'verified' && QA whiteboard contains 'Verified in upstream'`.
| infrastructure | workaround upstream bug closure in downstream jobs skip downstream tests if bug status verified qa whiteboard contains verified in upstream | 1 |
70,744 | 7,199,507,163 | IssuesEvent | 2018-02-05 16:08:50 | ubtue/tuefind | https://api.github.com/repos/ubtue/tuefind | closed | Anzeige der "Weiteren Schlagwörter" | ready for testing | Bitte einleitende "gnd" usw. aus der Anzeige entfernen

| 1.0 | Anzeige der "Weiteren Schlagwörter" - Bitte einleitende "gnd" usw. aus der Anzeige entfernen

| non_infrastructure | anzeige der weiteren schlagwörter bitte einleitende gnd usw aus der anzeige entfernen | 0 |
20,353 | 13,877,662,283 | IssuesEvent | 2020-10-17 05:22:53 | esrlabs/chipmunk | https://api.github.com/repos/esrlabs/chipmunk | opened | Only produce artifacts we need in gh actions | infrastructure | Some generated artifacts are not needed and should not end up as artifacts at all

| 1.0 | Only produce artifacts we need in gh actions - Some generated artifacts are not needed and should not end up as artifacts at all

| infrastructure | only produce artifacts we need in gh actions some generated artifacts are not needed and should not end up as artifacts at all | 1 |
83,092 | 3,626,320,370 | IssuesEvent | 2016-02-10 00:10:57 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | Make kubectl describers and resource printers more consistent | component/kubectl help-wanted kind/cleanup priority/P2 team/ux | Noticed when reviewing #20886
https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/describe.go
Describers should all print object metadata consistently (name, namespace if relevant, labels).
All controllers should print at least pod templates the same way -- using DescribePodTemplate.
I'm sure there are other opportunities to improve consistency. I just noticed these differences at a glance.
The resource printers (for get) are also inconsistent. @AdoHe has been improving on them (e.g., #20557, #20409).
https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/resource_printer.go
cc @kubernetes/kubectl @mqliang @erictune @mikedanese @janetkuo @kargakis @madhusudancs @AdoHe | 1.0 | Make kubectl describers and resource printers more consistent - Noticed when reviewing #20886
https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/describe.go
Describers should all print object metadata consistently (name, namespace if relevant, labels).
All controllers should print at least pod templates the same way -- using DescribePodTemplate.
I'm sure there are other opportunities to improve consistency. I just noticed these differences at a glance.
The resource printers (for get) are also inconsistent. @AdoHe has been improving on them (e.g., #20557, #20409).
https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/resource_printer.go
cc @kubernetes/kubectl @mqliang @erictune @mikedanese @janetkuo @kargakis @madhusudancs @AdoHe | non_infrastructure | make kubectl describers and resource printers more consistent noticed when reviewing describers should all print object metadata consistently name namespace if relevant labels all controllers should print at least pod templates the same way using describepodtemplate i m sure there are other opportunities to improve consistency i just noticed these differences at a glance the resource printers for get are also inconsistent adohe has been improving on them e g cc kubernetes kubectl mqliang erictune mikedanese janetkuo kargakis madhusudancs adohe | 0 |
13,279 | 10,180,241,540 | IssuesEvent | 2019-08-09 09:47:55 | raiden-network/raiden | https://api.github.com/repos/raiden-network/raiden | opened | Check if "Decreasing linting errors" script works | infrastructure | I've never seen the "Decreasing linting errors" step in CI fail. Either we're really good at increasing mypy errors, or the script doesn't work as expected. | 1.0 | Check if "Decreasing linting errors" script works - I've never seen the "Decreasing linting errors" step in CI fail. Either we're really good at increasing mypy errors, or the script doesn't work as expected. | infrastructure | check if decreasing linting errors script works i ve never seen the decreasing linting errors step in ci fail either we re really good at increasing mypy errors or the script doesn t work as expected | 1 |
3,884 | 4,691,431,734 | IssuesEvent | 2016-10-11 10:29:29 | elmsln/elmsln | https://api.github.com/repos/elmsln/elmsln | closed | During Amazon spin up. PHP Warning: Error while sending QUERY packet | bug Developer Experience infrastructure medium minor testing | PHP Warning: Error while sending QUERY packet. PID=18550 in /var/www/elmsln/core/dslmcode/shared/drupal-7.x/modules/ulmus/apdqc/apdqc.mysql.inc on line 860
Not sure what this is will investigate at some point... | 1.0 | During Amazon spin up. PHP Warning: Error while sending QUERY packet - PHP Warning: Error while sending QUERY packet. PID=18550 in /var/www/elmsln/core/dslmcode/shared/drupal-7.x/modules/ulmus/apdqc/apdqc.mysql.inc on line 860
Not sure what this is will investigate at some point... | infrastructure | during amazon spin up php warning error while sending query packet php warning error while sending query packet pid in var www elmsln core dslmcode shared drupal x modules ulmus apdqc apdqc mysql inc on line not sure what this is will investigate at some point | 1 |
26,354 | 20,022,491,403 | IssuesEvent | 2022-02-01 17:40:06 | cloud-native-toolkit/automation-modules | https://api.github.com/repos/cloud-native-toolkit/automation-modules | closed | Request new module: cp-truststore | new_module type:gitops category:infrastructure approved | **Related user story**
Provide the issue number for the user story related to this request - https://github.ibm.com/gsi-labs/industry-cloud-planning/issues/550
**Type of module**
gitops
**If requesting a terraform module, what is the target cloud platform?**
N/A
**What is the purpose of the module?**
Please describe what this automation module will do
**What is the category of the module?**
infrastructure
| 1.0 | Request new module: cp-truststore - **Related user story**
Provide the issue number for the user story related to this request - https://github.ibm.com/gsi-labs/industry-cloud-planning/issues/550
**Type of module**
gitops
**If requesting a terraform module, what is the target cloud platform?**
N/A
**What is the purpose of the module?**
Please describe what this automation module will do
**What is the category of the module?**
infrastructure
| infrastructure | request new module cp truststore related user story provide the issue number for the user story related to this request type of module gitops if requesting a terraform module what is the target cloud platform n a what is the purpose of the module please describe what this automation module will do what is the category of the module infrastructure | 1 |
27,627 | 22,045,308,504 | IssuesEvent | 2022-05-30 00:20:25 | dart-lang/site-www | https://api.github.com/repos/dart-lang/site-www | closed | Consider extracting link check from stage/deploy to catch link issues in PRs | enhancement infrastructure p2-medium e1-hours links | ### Describe the problem
The link check is tied to the build/stage jobs, which don't run in every user's PRs. We should consider extracting it from those jobs so it will catch link issues for all pull requests.
### Expected fix
Perhaps create a separate job to run the link checker. If possible avoid rebuilding the website again just for the link check.
### Additional context
_No response_ | 1.0 | Consider extracting link check from stage/deploy to catch link issues in PRs - ### Describe the problem
The link check is tied to the build/stage jobs, which don't run in every user's PRs. We should consider extracting it from those jobs so it will catch link issues for all pull requests.
### Expected fix
Perhaps create a separate job to run the link checker. If possible avoid rebuilding the website again just for the link check.
### Additional context
_No response_ | infrastructure | consider extracting link check from stage deploy to catch link issues in prs describe the problem the link check is tied to the build stage jobs which don t run in every user s prs we should consider extracting it from those jobs so it will catch link issues for all pull requests expected fix perhaps create a separate job to run the link checker if possible avoid rebuilding the website again just for the link check additional context no response | 1 |
25,830 | 19,273,655,903 | IssuesEvent | 2021-12-10 09:18:40 | chatwoot/chatwoot | https://api.github.com/repos/chatwoot/chatwoot | closed | dockerhub builds failing | infrastructure devops | **Describe the bug**
Dockerhub builds are failing since today morning.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://hub.docker.com/repository/docker/chatwoot/chatwoot/builds
**Expected behavior**
docker builds should pass
**Screenshots**
n/a
**Browser logs**
n/a
**Server logs**
n/a
**Environment**
n/a
**Desktop (please complete the following information):**
n/a
**Smartphone (please complete the following information):**
n/a
**Additional context**
n/a | 1.0 | dockerhub builds failing - **Describe the bug**
Dockerhub builds are failing since today morning.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://hub.docker.com/repository/docker/chatwoot/chatwoot/builds
**Expected behavior**
docker builds should pass
**Screenshots**
n/a
**Browser logs**
n/a
**Server logs**
n/a
**Environment**
n/a
**Desktop (please complete the following information):**
n/a
**Smartphone (please complete the following information):**
n/a
**Additional context**
n/a | infrastructure | dockerhub builds failing describe the bug dockerhub builds are failing since today morning to reproduce steps to reproduce the behavior go to expected behavior docker builds should pass screenshots n a browser logs n a server logs n a environment n a desktop please complete the following information n a smartphone please complete the following information n a additional context n a | 1 |
25,583 | 18,877,176,314 | IssuesEvent | 2021-11-14 07:20:40 | w3c/aria-practices | https://api.github.com/repos/w3c/aria-practices | closed | Remove dependency "npm-merge-driver" | Infrastructure | See #1766
When running `npm audit fix --force` it changes the version for `npm-merge-driver` to `^0.0.0`, which seems weird at first. Changing it back to `^2.3.6` and running `npm audit` again it finds 7 low severity vulnerabilities.
I'd like to get `npm audit` to report 0 vulnerabilities and keep it there. Until `npm-merge-driver` gets a new version, I suggest we drop this dependency. | 1.0 | Remove dependency "npm-merge-driver" - See #1766
When running `npm audit fix --force` it changes the version for `npm-merge-driver` to `^0.0.0`, which seems weird at first. Changing it back to `^2.3.6` and running `npm audit` again it finds 7 low severity vulnerabilities.
I'd like to get `npm audit` to report 0 vulnerabilities and keep it there. Until `npm-merge-driver` gets a new version, I suggest we drop this dependency. | infrastructure | remove dependency npm merge driver see when running npm audit fix force it changes the version for npm merge driver to which seems weird at first changing it back to and running npm audit again it finds low severity vulnerabilities i d like to get npm audit to report vulnerabilities and keep it there until npm merge driver gets a new version i suggest we drop this dependency | 1 |
21,037 | 14,286,729,056 | IssuesEvent | 2020-11-23 15:30:02 | sunpy/sunpy | https://api.github.com/repos/sunpy/sunpy | closed | Intermittent image-rotation test failures on OS X when using conda | Effort High Infrastructure Package Expert Priority High Tests image | We have intermittent failures of our image rotation tests and these failures appear to be isolated to OS X when using conda(-forge). #4235 added raw output for these failures occurred, and some of the output is truly bizarre. I will add investigative stuff in separate posts. My current conjecture is there is nothing wrong with SunPy code, but rather a C extension in the `numpy`/`scipy`/`scikit-image` ecosystem is not being compiled for conda(-forge) with the correct compile options for OS X such that there's the intermittent potential for bad memory access of arrays.
---
Edit: go down to https://github.com/sunpy/sunpy/issues/4290#issuecomment-676573472 for a summary of the current understanding | 1.0 | Intermittent image-rotation test failures on OS X when using conda - We have intermittent failures of our image rotation tests and these failures appear to be isolated to OS X when using conda(-forge). #4235 added raw output for these failures occurred, and some of the output is truly bizarre. I will add investigative stuff in separate posts. My current conjecture is there is nothing wrong with SunPy code, but rather a C extension in the `numpy`/`scipy`/`scikit-image` ecosystem is not being compiled for conda(-forge) with the correct compile options for OS X such that there's the intermittent potential for bad memory access of arrays.
---
Edit: go down to https://github.com/sunpy/sunpy/issues/4290#issuecomment-676573472 for a summary of the current understanding | infrastructure | intermittent image rotation test failures on os x when using conda we have intermittent failures of our image rotation tests and these failures appear to be isolated to os x when using conda forge added raw output for these failures occurred and some of the output is truly bizarre i will add investigative stuff in separate posts my current conjecture is there is nothing wrong with sunpy code but rather a c extension in the numpy scipy scikit image ecosystem is not being compiled for conda forge with the correct compile options for os x such that there s the intermittent potential for bad memory access of arrays edit go down to for a summary of the current understanding | 1 |
17,319 | 12,300,903,737 | IssuesEvent | 2020-05-11 14:39:31 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | closed | Maintenance site | Component: Frontend Component: Infrastructure Difficulty: Medium Priority: High Status: In Progress Type: Enhancement | When deploying changes for the TM now with the frontend being separated from the backend. The frontend usually persists, while the backend is not reachable. For the user it looks like the tool is working well, but it doesn't behave well. Lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors. | 1.0 | Maintenance site - When deploying changes for the TM now with the frontend being separated from the backend. The frontend usually persists, while the backend is not reachable. For the user it looks like the tool is working well, but it doesn't behave well. Lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors. | infrastructure | maintenance site when deploying changes for the tm now with the frontend being separated from the backend the frontend usually persists while the backend is not reachable for the user it looks like the tool is working well but it doesn t behave well lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors | 1 |
34,964 | 30,635,149,049 | IssuesEvent | 2023-07-24 17:10:55 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Wasm.Build.Tests.Blazor.MiscTests.NativeBuild_WithDeployOnBuild_UsedByVS failing in CI due to SIGKILL | arch-wasm area-Infrastructure-mono | e.g.
```
[] Linking with emcc with -O0. This may take a while ...
[] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-ld" @/datadisks/disk1/work/B6EC09B5/t/emscripten_3kui486x.rsp.utf-8
[] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize" -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features
[] emcc : error : '/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features' failed (received SIGKILL (-9)) [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj]
[] /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.WebAssembly.Sdk/8.0.0-ci/Sdk/WasmApp.Native.targets(431,5): error MSB3073: The command "emcc "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-default.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-link.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/emcc-link.rsp"" exited with code 1. [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj]
```
https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-78580-merge-c362e9e2c2db48c69b/Workloads-Wasm.Build.Tests.Blazor.MiscTests/1/console.d18bdf9a.log?helixlogtype=result | 1.0 | Wasm.Build.Tests.Blazor.MiscTests.NativeBuild_WithDeployOnBuild_UsedByVS failing in CI due to SIGKILL - e.g.
```
[] Linking with emcc with -O0. This may take a while ...
[] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-ld" @/datadisks/disk1/work/B6EC09B5/t/emscripten_3kui486x.rsp.utf-8
[] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize" -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features
[] emcc : error : '/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features' failed (received SIGKILL (-9)) [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj]
[] /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.WebAssembly.Sdk/8.0.0-ci/Sdk/WasmApp.Native.targets(431,5): error MSB3073: The command "emcc "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-default.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-link.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/emcc-link.rsp"" exited with code 1. [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj]
```
https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-78580-merge-c362e9e2c2db48c69b/Workloads-Wasm.Build.Tests.Blazor.MiscTests/1/console.d18bdf9a.log?helixlogtype=result | infrastructure | wasm build tests blazor misctests nativebuild withdeployonbuild usedbyvs failing in ci due to sigkill e g linking with emcc with this may take a while datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm ld datadisks work t emscripten rsp utf datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm emscripten finalize g dyncalls dwarf datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm o datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm detect features emcc error datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm emscripten finalize g dyncalls dwarf datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm o datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm detect features failed received sigkill datadisks work w e dotnet latest packs microsoft net runtime webassembly sdk ci sdk wasmapp native targets error the command emcc datadisks work w e dotnet latest packs microsoft netcore app runtime mono browser wasm ci runtimes browser wasm native src emcc default rsp datadisks work w e dotnet latest packs microsoft netcore app runtime mono browser wasm ci runtimes browser wasm native src emcc link rsp datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish emcc link rsp exited with code | 1 |
241,737 | 18,473,005,416 | IssuesEvent | 2021-10-18 01:28:23 | edgexfoundry/edgex-docs | https://api.github.com/repos/edgexfoundry/edgex-docs | closed | Listing of supported device protocols | documentation enhancement help wanted good first issue 1-low jakarta | We need some documentation on the supported device services and protocols. In order not to have to keep updating the page with each new DS added, it would be ok to show how to find which device services we support (for example through GitHub).
This question comes up a lot in Slack and via email forums. Need a good place to send people to see the list. | 1.0 | Listing of supported device protocols - We need some documentation on the supported device services and protocols. In order not to have to keep updating the page with each new DS added, it would be ok to show how to find which device services we support (for example through GitHub).
This question comes up a lot in Slack and via email forums. Need a good place to send people to see the list. | non_infrastructure | listing of supported device protocols we need some documentation on the supported device services and protocols in order not to have to keep updating the page with each new ds added it would be ok to show how to find which device services we support for example through github this question comes up a lot in slack and via email forums need a good place to send people to see the list | 0 |
10,856 | 8,759,715,714 | IssuesEvent | 2018-12-15 19:14:00 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | Optimize performance of push notifications code in test suite | area: notifications (messages) area: testing-infrastructure good first issue help wanted in progress | If you add a `return` statement at the top of `handle_push_notifications`, running `test-backend test_notifications` goes from taking 4.3s to 3.6s on my laptop. I suspect there are other tests where we could save significant time as well, but it's hard to run the full test suite that way, because `handle_push_notifications` does useful stuff.
What I'm thinking is that we can/should add an early-return in that function is `push_notification_enabled()` is False, and see if that helps improve the test suite performance.
This issue doesn't require a ton of Zulip background, but it may require some good Python test system debugging skills (the tests that are likely to fail as one works on this involve a decent amount of mocks).
@HarshitOnGitHub this may be an issue of interest for you when you have time again if nobody else does it first. | 1.0 | Optimize performance of push notifications code in test suite - If you add a `return` statement at the top of `handle_push_notifications`, running `test-backend test_notifications` goes from taking 4.3s to 3.6s on my laptop. I suspect there are other tests where we could save significant time as well, but it's hard to run the full test suite that way, because `handle_push_notifications` does useful stuff.
What I'm thinking is that we can/should add an early-return in that function is `push_notification_enabled()` is False, and see if that helps improve the test suite performance.
This issue doesn't require a ton of Zulip background, but it may require some good Python test system debugging skills (the tests that are likely to fail as one works on this involve a decent amount of mocks).
@HarshitOnGitHub this may be an issue of interest for you when you have time again if nobody else does it first. | infrastructure | optimize performance of push notifications code in test suite if you add a return statement at the top of handle push notifications running test backend test notifications goes from taking to on my laptop i suspect there are other tests where we could save significant time as well but it s hard to run the full test suite that way because handle push notifications does useful stuff what i m thinking is that we can should add an early return in that function is push notification enabled is false and see if that helps improve the test suite performance this issue doesn t require a ton of zulip background but it may require some good python test system debugging skills the tests that are likely to fail as one works on this involve a decent amount of mocks harshitongithub this may be an issue of interest for you when you have time again if nobody else does it first | 1 |
297,831 | 22,405,829,672 | IssuesEvent | 2022-06-18 00:56:59 | Equipment-and-Tool-Institute/j1939-84 | https://api.github.com/repos/Equipment-and-Tool-Institute/j1939-84 | closed | Create and Affirm A-7 Tables for Minimum Test Results Support | documentation | Update A-7 Tables For 2024 MY Effects
Create with user community and review with CARB.
Table A-7 provides minimum test results content for DM4 and DM7.
Continuous tests for Misfire and Fuel Quantity obviate discrete test results for DM7.
2019MY+ with continuous monitors are failing.
See #1125 for posted background on misfire. | 1.0 | Create and Affirm A-7 Tables for Minimum Test Results Support - Update A-7 Tables For 2024 MY Effects
Create with user community and review with CARB.
Table A-7 provides minimum test results content for DM4 and DM7.
Continuous tests for Misfire and Fuel Quantity obviate discrete test results for DM7.
2019MY+ with continuous monitors are failing.
See #1125 for posted background on misfire. | non_infrastructure | create and affirm a tables for minimum test results support update a tables for my effects create with user community and review with carb table a provides minimum test results content for and continuous tests for misfire and fuel quantity obviate discrete test results for with continuous monitors are failing see for posted background on misfire | 0 |
183,715 | 21,781,981,599 | IssuesEvent | 2022-05-13 20:04:30 | kira0x1/mikaela | https://api.github.com/repos/kira0x1/mikaela | closed | discord.js-12.3.1.tgz: 1 vulnerabilities (highest severity is: 6.1) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>discord.js-12.3.1.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-0235](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | node-fetch-2.6.1.tgz | Transitive | 12.5.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0235</summary>
### Vulnerable Library - <b>node-fetch-2.6.1.tgz</b></p>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- discord.js-12.3.1.tgz (Root Library)
- :x: **node-fetch-2.6.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor
<p>Publish Date: 2022-01-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235>CVE-2022-0235</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-r683-j2x4-v87g">https://github.com/advisories/GHSA-r683-j2x4-v87g</a></p>
<p>Release Date: 2022-01-16</p>
<p>Fix Resolution (node-fetch): 2.6.7</p>
<p>Direct dependency fix Resolution (discord.js): 12.5.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"discord.js","packageVersion":"12.3.1","packageFilePaths":["/tmp/ws-scm/MikaelaBot/package.json"],"isTransitiveDependency":false,"dependencyTree":"discord.js:12.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"12.5.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-0235","vulnerabilityDetails":"node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | True | discord.js-12.3.1.tgz: 1 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>discord.js-12.3.1.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-0235](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | node-fetch-2.6.1.tgz | Transitive | 12.5.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0235</summary>
### Vulnerable Library - <b>node-fetch-2.6.1.tgz</b></p>
<p>A light-weight module that brings window.fetch to node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- discord.js-12.3.1.tgz (Root Library)
- :x: **node-fetch-2.6.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor
<p>Publish Date: 2022-01-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235>CVE-2022-0235</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-r683-j2x4-v87g">https://github.com/advisories/GHSA-r683-j2x4-v87g</a></p>
<p>Release Date: 2022-01-16</p>
<p>Fix Resolution (node-fetch): 2.6.7</p>
<p>Direct dependency fix Resolution (discord.js): 12.5.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"discord.js","packageVersion":"12.3.1","packageFilePaths":["/tmp/ws-scm/MikaelaBot/package.json"],"isTransitiveDependency":false,"dependencyTree":"discord.js:12.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"12.5.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-0235","vulnerabilityDetails":"node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | non_infrastructure | discord js tgz vulnerabilities highest severity is vulnerable library discord js tgz path to dependency file tmp ws scm mikaelabot package json path to vulnerable library tmp ws scm mikaelabot node modules node fetch package json vulnerabilities cve severity cvss dependency type fixed in remediation available medium node fetch tgz transitive details cve vulnerable library node fetch tgz a light weight module that brings window fetch to node js library home page a href path to dependency file tmp ws scm mikaelabot package json path to vulnerable library tmp ws scm mikaelabot node modules node fetch package json dependency hierarchy discord js tgz root library x node fetch tgz vulnerable library found in base branch main vulnerability details node fetch is vulnerable to exposure of sensitive information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node fetch direct dependency fix resolution discord js step up your open source security game with whitesource istransitivedependency false dependencytree discord js isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node fetch is vulnerable to exposure of sensitive information to an unauthorized actor vulnerabilityurl | 0 |
255,019 | 8,102,473,709 | IssuesEvent | 2018-08-13 01:58:25 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | closed | Deleting error-report not replicating properly | Priority: 2 - Medium Status: 5 - Ready Type: Bug UI/UX | I deleted a report in couchdb while still open in the UI. Back to UI, the LHS still shows the report…
**Steps to reproduce**:
-_ Login `alpha.dev` as admin (or other instance)_
- _Create a report via sms with an invalid data: eg send sms (using admin console or otherwise): `1!C!125#Romeo` _
-_Notice that the report appears on the LHS in the UI

_
-_Click on the report and notice is opens in the RHS with correct error message_

_copy the report uuid, delete the doc in Futon (or maybe other non-UI ways to delete the report)_
_Look back at the UI_
**What should happen**:
- _the report should be completely deleted from the UI_
-_The browser should redirect to `/_design/medic/_rewrite/#/reports/`_
**What actually happens**:
- _The report still shows in the LHS (until a manual refresh)_
-_The url still has the report id_

- _Error in console_

**Environment**:
_Fill in as many of these as apply. It's especially useful if some environments work and some don't (ie works in Chrome but not in FF)._
- Instance: _(tested alpha.dev.medicmobile.org, and beta-old)_
- Browser: _(Firefox, Chrome)_
- Client platform: _(MacOS)_
- App: _(webapp)_
- Version: _(2.16.1-beta.1, 3.1.0)_
**Other**:
_The UI `delete` works as expected in this case_
| 1.0 | Deleting error-report not replicating properly - I deleted a report in couchdb while still open in the UI. Back to UI, the LHS still shows the report…
**Steps to reproduce**:
-_ Login `alpha.dev` as admin (or other instance)_
- _Create a report via sms with an invalid data: eg send sms (using admin console or otherwise): `1!C!125#Romeo` _
-_Notice that the report appears on the LHS in the UI

_
-_Click on the report and notice is opens in the RHS with correct error message_

_copy the report uuid, delete the doc in Futon (or maybe other non-UI ways to delete the report)_
_Look back at the UI_
**What should happen**:
- _the report should be completely deleted from the UI_
-_The browser should redirect to `/_design/medic/_rewrite/#/reports/`_
**What actually happens**:
- _The report still shows in the LHS (until a manual refresh)_
-_The url still has the report id_

- _Error in console_

**Environment**:
_Fill in as many of these as apply. It's especially useful if some environments work and some don't (ie works in Chrome but not in FF)._
- Instance: _(tested alpha.dev.medicmobile.org, and beta-old)_
- Browser: _(Firefox, Chrome)_
- Client platform: _(MacOS)_
- App: _(webapp)_
- Version: _(2.16.1-beta.1, 3.1.0)_
**Other**:
_The UI `delete` works as expected in this case_
| non_infrastructure | deleting error report not replicating properly i deleted a report in couchdb while still open in the ui back to ui the lhs still shows the report… steps to reproduce login alpha dev as admin or other instance create a report via sms with an invalid data eg send sms using admin console or otherwise c romeo notice that the report appears on the lhs in the ui click on the report and notice is opens in the rhs with correct error message copy the report uuid delete the doc in futon or maybe other non ui ways to delete the report look back at the ui what should happen the report should be completely deleted from the ui the browser should redirect to design medic rewrite reports what actually happens the report still shows in the lhs until a manual refresh the url still has the report id error in console environment fill in as many of these as apply it s especially useful if some environments work and some don t ie works in chrome but not in ff instance tested alpha dev medicmobile org and beta old browser firefox chrome client platform macos app webapp version beta other the ui delete works as expected in this case | 0 |
87,341 | 3,750,058,792 | IssuesEvent | 2016-03-11 03:46:09 | GLolol/PyLink | https://api.github.com/repos/GLolol/PyLink | opened | opercmds: 'trace'/'massban' command | feature priority:wishlist | Similar to Atheme contrib's `os_trace` module, this would be something that can match users by various criteria (`nick!user@host` globs or targets mentioned in #170) and then perform actions on them: `KILL`, `KLINE/ZLINE` (#139), `LIST`, etc. | 1.0 | opercmds: 'trace'/'massban' command - Similar to Atheme contrib's `os_trace` module, this would be something that can match users by various criteria (`nick!user@host` globs or targets mentioned in #170) and then perform actions on them: `KILL`, `KLINE/ZLINE` (#139), `LIST`, etc. | non_infrastructure | opercmds trace massban command similar to atheme contrib s os trace module this would be something that can match users by various criteria nick user host globs or targets mentioned in and then perform actions on them kill kline zline list etc | 0 |
19,472 | 5,889,452,962 | IssuesEvent | 2017-05-17 12:56:51 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Menu reset in filter after creating new menu item | No Code Attached Yet | ### Steps to reproduce the issue
I have a menu with about 40 items. I go to Menus > My menu and see list of items:

After creating new menu item I returning to the list of all items and no menu is selected:

That's very comfortless when you create mass menu items.
### Expected result
After creating new menu item we must be returned to the same menu.
### Actual result
I returning to All menu items.
### System information (as much as possible)
Joomla 3.7.0
PHP Version 7.0.17 | 1.0 | Menu reset in filter after creating new menu item - ### Steps to reproduce the issue
I have a menu with about 40 items. I go to Menus > My menu and see list of items:

After creating new menu item I returning to the list of all items and no menu is selected:

That's very comfortless when you create mass menu items.
### Expected result
After creating new menu item we must be returned to the same menu.
### Actual result
I returning to All menu items.
### System information (as much as possible)
Joomla 3.7.0
PHP Version 7.0.17 | non_infrastructure | menu reset in filter after creating new menu item steps to reproduce the issue i have a menu with about items i go to menus my menu and see list of items after creating new menu item i returning to the list of all items and no menu is selected that s very comfortless when you create mass menu items expected result after creating new menu item we must be returned to the same menu actual result i returning to all menu items system information as much as possible joomla php version | 0 |
160,051 | 12,501,029,276 | IssuesEvent | 2020-06-02 00:00:09 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | opened | Flaky test PublicExposureOfPubternalTypeProducesPUB0001 | test-investigation | https://dev.azure.com/dnceng/public/_build/results?buildId=663272&view=results
PublicExposureOfPubternalTypeProducesPUB0001
```
[xUnit.net 00:00:08.07] Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(member: "public /*MM*/CD c { get; }") [FAIL]
[xUnit.net 00:00:08.08] The collection was expected to contain a single element, but it was empty.
[xUnit.net 00:00:08.08] Stack Trace:
[xUnit.net 00:00:08.08] /_/src/Analyzers/Internal.AspNetCore.Analyzers/test/PubternabilityAnalyzerTests.cs(47,0): at Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(String member)
[xUnit.net 00:00:08.08] --- End of stack trace from previous location ---
[xUnit.net 00:00:08.08] Output:
[xUnit.net 00:00:08.08] Adding file: Test0
[xUnit.net 00:00:08.08] using A.Internal.Namespace;
[xUnit.net 00:00:08.08] namespace A.Internal.Namespace
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public class C {}
[xUnit.net 00:00:08.08] public delegate C CD ();
[xUnit.net 00:00:08.08] public class CAAttribute: System.Attribute {}
[xUnit.net 00:00:08.08]
[xUnit.net 00:00:08.08] public class Program
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public static void Main() {}
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] namespace A
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public class T
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public CD c { get; }
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] }
``` | 1.0 | Flaky test PublicExposureOfPubternalTypeProducesPUB0001 - https://dev.azure.com/dnceng/public/_build/results?buildId=663272&view=results
PublicExposureOfPubternalTypeProducesPUB0001
```
[xUnit.net 00:00:08.07] Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(member: "public /*MM*/CD c { get; }") [FAIL]
[xUnit.net 00:00:08.08] The collection was expected to contain a single element, but it was empty.
[xUnit.net 00:00:08.08] Stack Trace:
[xUnit.net 00:00:08.08] /_/src/Analyzers/Internal.AspNetCore.Analyzers/test/PubternabilityAnalyzerTests.cs(47,0): at Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(String member)
[xUnit.net 00:00:08.08] --- End of stack trace from previous location ---
[xUnit.net 00:00:08.08] Output:
[xUnit.net 00:00:08.08] Adding file: Test0
[xUnit.net 00:00:08.08] using A.Internal.Namespace;
[xUnit.net 00:00:08.08] namespace A.Internal.Namespace
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public class C {}
[xUnit.net 00:00:08.08] public delegate C CD ();
[xUnit.net 00:00:08.08] public class CAAttribute: System.Attribute {}
[xUnit.net 00:00:08.08]
[xUnit.net 00:00:08.08] public class Program
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public static void Main() {}
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] namespace A
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public class T
[xUnit.net 00:00:08.08] {
[xUnit.net 00:00:08.08] public CD c { get; }
[xUnit.net 00:00:08.08] }
[xUnit.net 00:00:08.08] }
``` | non_infrastructure | flaky test internal aspnetcore analyzers tests pubternabilityanalyzertests member public mm cd c get the collection was expected to contain a single element but it was empty stack trace src analyzers internal aspnetcore analyzers test pubternabilityanalyzertests cs at internal aspnetcore analyzers tests pubternabilityanalyzertests string member end of stack trace from previous location output adding file using a internal namespace namespace a internal namespace public class c public delegate c cd public class caattribute system attribute public class program public static void main namespace a public class t public cd c get | 0 |
2,228 | 3,576,276,087 | IssuesEvent | 2016-01-27 18:55:59 | dotnet/roslyn-analyzers | https://api.github.com/repos/dotnet/roslyn-analyzers | closed | "Project Site" link on NuGet.org points to wrong site. | Area-Infrastructure Bug | For example, the "Project Site" link on https://www.nuget.org/packages/System.Runtime.InteropServices.Analyzers/ points to http://msdn.com/roslyn, which redirects to http://github.com/dotnet/roslyn.
We should be directing users back to http://github.com/dotnet/roslyn-analyzers instead. | 1.0 | "Project Site" link on NuGet.org points to wrong site. - For example, the "Project Site" link on https://www.nuget.org/packages/System.Runtime.InteropServices.Analyzers/ points to http://msdn.com/roslyn, which redirects to http://github.com/dotnet/roslyn.
We should be directing users back to http://github.com/dotnet/roslyn-analyzers instead. | infrastructure | project site link on nuget org points to wrong site for example the project site link on points to which redirects to we should be directing users back to instead | 1 |
28,032 | 22,776,041,889 | IssuesEvent | 2022-07-08 14:33:17 | Zilliqa/scilla | https://api.github.com/repos/Zilliqa/scilla | opened | Move to Ubuntu 20.04 | infrastructure | Things that we need to change for this:
- [ ] Docker and Docker.slim files
- [ ] Make sure Travis CI uses Ubuntu 20.04
- [ ] Change `INSTALL.md` file accordingly (including the WSL installation) | 1.0 | Move to Ubuntu 20.04 - Things that we need to change for this:
- [ ] Docker and Docker.slim files
- [ ] Make sure Travis CI uses Ubuntu 20.04
- [ ] Change `INSTALL.md` file accordingly (including the WSL installation) | infrastructure | move to ubuntu things that we need to change for this docker and docker slim files make sure travis ci uses ubuntu change install md file accordingly including the wsl installation | 1 |
30,986 | 25,225,542,303 | IssuesEvent | 2022-11-14 15:43:35 | SonarSource/sonar-scanner-msbuild | https://api.github.com/repos/SonarSource/sonar-scanner-msbuild | closed | Fix broken ITs: Change tested plugin versions | Infrastructure | Integration tests uses LATEST_RELEASE of plugins with old SQ instances.
The assumption that it should work is not valid anymore since the release strategy changed few years ago.
We need to update Orchestration to:
* use plugins that were shipped with the given version of SQ
* add IT with LATEST_RELEASE of SQ and use analyzer LATEST_RELEASE only there | 1.0 | Fix broken ITs: Change tested plugin versions - Integration tests uses LATEST_RELEASE of plugins with old SQ instances.
The assumption that it should work is not valid anymore since the release strategy changed few years ago.
We need to update Orchestration to:
* use plugins that were shipped with the given version of SQ
* add IT with LATEST_RELEASE of SQ and use analyzer LATEST_RELEASE only there | infrastructure | fix broken its change tested plugin versions integration tests uses latest release of plugins with old sq instances the assumption that it should work is not valid anymore since the release strategy changed few years ago we need to update orchestration to use plugins that were shipped with the given version of sq add it with latest release of sq and use analyzer latest release only there | 1 |
56,186 | 13,770,942,545 | IssuesEvent | 2020-10-07 21:04:58 | GoogleCloudPlatform/ruby-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/ruby-docs-samples | closed | Minitest::Result: test_0009_in_query_without_array failed | :rotating_light: api: firestore buildcop: flaky buildcop: issue priority: p1 samples type: bug | This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: d52320406beaa73ab9b4e668d742acac43f33461
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/23be4cb1-8456-4efd-a366-3b5d38760de8), [Sponge](http://sponge2/23be4cb1-8456-4efd-a366-3b5d38760de8)
status: failed
<details><summary>Test output</summary><br><pre>/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:183:in `assert'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:264:in `assert_includes'
/tmpfs/src/github/ruby-docs-samples/firestore/acceptance/query_data_test.rb:121:in `block (2 levels) in <top (required)>'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:98:in `block (3 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:195:in `capture_exceptions'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:95:in `block (2 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:272:in `time_it'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:94:in `block in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:211:in `with_info_handler'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:93:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:1029:in `run_one_method'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:341:in `run_one_method'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:328:in `block (2 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `each'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `block in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:354:in `with_info_handler'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:326:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `block in __run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `map'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `__run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:141:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:68:in `block in autorun'</pre></details> | 2.0 | Minitest::Result: test_0009_in_query_without_array failed - This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: d52320406beaa73ab9b4e668d742acac43f33461
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/23be4cb1-8456-4efd-a366-3b5d38760de8), [Sponge](http://sponge2/23be4cb1-8456-4efd-a366-3b5d38760de8)
status: failed
<details><summary>Test output</summary><br><pre>/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:183:in `assert'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:264:in `assert_includes'
/tmpfs/src/github/ruby-docs-samples/firestore/acceptance/query_data_test.rb:121:in `block (2 levels) in <top (required)>'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:98:in `block (3 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:195:in `capture_exceptions'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:95:in `block (2 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:272:in `time_it'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:94:in `block in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:211:in `with_info_handler'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:93:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:1029:in `run_one_method'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:341:in `run_one_method'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:328:in `block (2 levels) in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `each'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `block in run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:354:in `with_info_handler'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:326:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `block in __run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `map'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `__run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:141:in `run'
/usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:68:in `block in autorun'</pre></details> | non_infrastructure | minitest result test in query without array failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output usr local bundle gems minitest lib minitest assertions rb in assert usr local bundle gems minitest lib minitest assertions rb in assert includes tmpfs src github ruby docs samples firestore acceptance query data test rb in block levels in usr local bundle gems minitest lib minitest test rb in block levels in run usr local bundle gems minitest lib minitest test rb in capture exceptions usr local bundle gems minitest lib minitest test rb in block levels in run usr local bundle gems minitest lib minitest rb in time it usr local bundle gems minitest lib minitest test rb in block in run usr local bundle gems minitest lib minitest rb in on signal usr local bundle gems minitest lib minitest test rb in with info handler usr local bundle gems minitest lib minitest test rb in run usr local bundle gems minitest lib minitest rb in run one method usr local bundle gems minitest lib minitest rb in run one method usr local bundle gems minitest lib minitest rb in block levels in run usr local bundle gems minitest lib minitest rb in each usr local bundle gems minitest lib minitest rb in block in run usr local bundle gems minitest lib minitest rb in on signal usr local bundle gems minitest lib minitest rb in with info handler usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in block in run usr local bundle gems minitest lib minitest rb in map usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in block in autorun | 0 |
538,502 | 15,770,299,488 | IssuesEvent | 2021-03-31 19:17:03 | ArchitectSMP/k8s-gitops | https://api.github.com/repos/ArchitectSMP/k8s-gitops | closed | Proxy Service for improving American connection to server | enhancement feature/cluster priority/medium | Look into Proxy Service for improving American connection to server | 1.0 | Proxy Service for improving American connection to server - Look into Proxy Service for improving American connection to server | non_infrastructure | proxy service for improving american connection to server look into proxy service for improving american connection to server | 0 |
48,013 | 5,931,076,637 | IssuesEvent | 2017-05-24 04:35:58 | pump-io/pump.io | https://api.github.com/repos/pump-io/pump.io | opened | Make sure multiple tests don't have the same title | test suite | I just went to debug "User stream API test" and discovered that there are _three tests_ with this suite description. This seems really unfortunate. | 1.0 | Make sure multiple tests don't have the same title - I just went to debug "User stream API test" and discovered that there are _three tests_ with this suite description. This seems really unfortunate. | non_infrastructure | make sure multiple tests don t have the same title i just went to debug user stream api test and discovered that there are three tests with this suite description this seems really unfortunate | 0 |
110,851 | 13,943,790,536 | IssuesEvent | 2020-10-23 00:06:04 | 1aurabrown/o2-monde-shopify | https://api.github.com/repos/1aurabrown/o2-monde-shopify | closed | Collection Page Re-Design | design | - [x] small text
- [x] text blocks
- similar to homepage quotes
- [x] pagination
- [x] product card:
- [x] style on top center
- [x] color just below
- [x] price and sizes bottom
- [x] option with "size" label
- [x] text appears on hover
- [x] preorder tag
Mock Up States:
- [x] sale
- [x] some sizes sold out
- [x] hover
- [x] all sizes sold out
- [x] sizes label
- [x] 4px grid gutter option | 1.0 | Collection Page Re-Design - - [x] small text
- [x] text blocks
- similar to homepage quotes
- [x] pagination
- [x] product card:
- [x] style on top center
- [x] color just below
- [x] price and sizes bottom
- [x] option with "size" label
- [x] text appears on hover
- [x] preorder tag
Mock Up States:
- [x] sale
- [x] some sizes sold out
- [x] hover
- [x] all sizes sold out
- [x] sizes label
- [x] 4px grid gutter option | non_infrastructure | collection page re design small text text blocks similar to homepage quotes pagination product card style on top center color just below price and sizes bottom option with size label text appears on hover preorder tag mock up states sale some sizes sold out hover all sizes sold out sizes label grid gutter option | 0 |
29,116 | 23,729,578,497 | IssuesEvent | 2022-08-30 23:40:39 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | WorkspacesNetCore.MetadataReference should not need to wait for all asynchronous operations | Bug Area-Infrastructure Integration-Test | **Version Used**: #34265
The following wait operation should not be required:
https://github.com/dotnet/roslyn/blob/8c5af921afd4fc1a4e949052571e6503b0ca8288/src/VisualStudio/IntegrationTest/IntegrationTests/Workspace/WorkspacesNetCore.cs#L42-L44
It should only need to wait for `FeatureAttribute.Workspace`. | 1.0 | WorkspacesNetCore.MetadataReference should not need to wait for all asynchronous operations - **Version Used**: #34265
The following wait operation should not be required:
https://github.com/dotnet/roslyn/blob/8c5af921afd4fc1a4e949052571e6503b0ca8288/src/VisualStudio/IntegrationTest/IntegrationTests/Workspace/WorkspacesNetCore.cs#L42-L44
It should only need to wait for `FeatureAttribute.Workspace`. | infrastructure | workspacesnetcore metadatareference should not need to wait for all asynchronous operations version used the following wait operation should not be required it should only need to wait for featureattribute workspace | 1 |
24,732 | 17,662,597,104 | IssuesEvent | 2021-08-21 20:32:25 | pathfinder-for-autonomous-navigation/FlightSoftware | https://api.github.com/repos/pathfinder-for-autonomous-navigation/FlightSoftware | opened | CI Timing Out Due to Poor PTest Performance | bug ptest infrastructure | TL;DR; PTest sucks and isn't shutting down the flight software binary processes properly. Running all the mission checkouts locally on my desktop eventually pinned all 24 cores and grabbed 32 gigs of RAM plus 38 of swap crashing other processes running in the background -- like my chrome tabs.
Upgrade PTest to properly shut things down. | 1.0 | CI Timing Out Due to Poor PTest Performance - TL;DR; PTest sucks and isn't shutting down the flight software binary processes properly. Running all the mission checkouts locally on my desktop eventually pinned all 24 cores and grabbed 32 gigs of RAM plus 38 of swap crashing other processes running in the background -- like my chrome tabs.
Upgrade PTest to properly shut things down. | infrastructure | ci timing out due to poor ptest performance tl dr ptest sucks and isn t shutting down the flight software binary processes properly running all the mission checkouts locally on my desktop eventually pinned all cores and grabbed gigs of ram plus of swap crashing other processes running in the background like my chrome tabs upgrade ptest to properly shut things down | 1 |
293,926 | 9,011,172,812 | IssuesEvent | 2019-02-05 14:04:19 | raster-foundry/raster-foundry-python-client | https://api.github.com/repos/raster-foundry/raster-foundry-python-client | opened | Add example of creating uploads from existing S3 bucket | client requested priority | A lot of times we have a set of files in S3 (hundreds or thousands) that we want to import, but there isn't a great way or an example for how to do this. | 1.0 | Add example of creating uploads from existing S3 bucket - A lot of times we have a set of files in S3 (hundreds or thousands) that we want to import, but there isn't a great way or an example for how to do this. | non_infrastructure | add example of creating uploads from existing bucket a lot of times we have a set of files in hundreds or thousands that we want to import but there isn t a great way or an example for how to do this | 0 |
20,515 | 13,961,869,926 | IssuesEvent | 2020-10-25 06:17:53 | SirJosh3917/smiley-face-game | https://api.github.com/repos/SirJosh3917/smiley-face-game | opened | Allow users to change their password | client enhancement infrastructure server ui | This would require
- Way to send emails to the afflicted account email
- Generate short-living JWT specifically for resetting a password
- UI for clicking "reset my password", and UI for resetting the password | 1.0 | Allow users to change their password - This would require
- Way to send emails to the afflicted account email
- Generate short-living JWT specifically for resetting a password
- UI for clicking "reset my password", and UI for resetting the password | infrastructure | allow users to change their password this would require way to send emails to the afflicted account email generate short living jwt specifically for resetting a password ui for clicking reset my password and ui for resetting the password | 1 |
9,186 | 7,858,093,440 | IssuesEvent | 2018-06-21 12:59:31 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | serve/warns_on_assets_paths_test is failing on the Windows bot, but not locally | Type: bug area-infrastructure | I just checked in a new test for pub. It passes everywhere, including on my local Windows machine. But it's failing on the bot with:
Results of running 'pub serve':
Process exited with exit code -1.
Standard output:
| MSG : Serving myapp on http://127.0.0.1:59461
| MSG : Build completed successfully
Standard error:
| FINE: Pub 0.1.2+3
| IO : Spawning cmd /c ver
| IO : Finished ver. Exit code 0.stdout:Microsoft Windows [Version 6.1.7601]Nothing output on stderr.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo\assets.
I didn't think a negative exit code was even possible. As you can see, no exception is thrown. It looks like the process just dies.
http://chromegw.corp.google.com/i/client.dart/builders/pub-win-be/builds/2695/steps/pub%20tests/logs/stdio
Any idea what's going on here?
| 1.0 | serve/warns_on_assets_paths_test is failing on the Windows bot, but not locally - I just checked in a new test for pub. It passes everywhere, including on my local Windows machine. But it's failing on the bot with:
Results of running 'pub serve':
Process exited with exit code -1.
Standard output:
| MSG : Serving myapp on http://127.0.0.1:59461
| MSG : Build completed successfully
Standard error:
| FINE: Pub 0.1.2+3
| IO : Spawning cmd /c ver
| IO : Finished ver. Exit code 0.stdout:Microsoft Windows [Version 6.1.7601]Nothing output on stderr.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo.
| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo\assets.
I didn't think a negative exit code was even possible. As you can see, no exception is thrown. It looks like the process just dies.
http://chromegw.corp.google.com/i/client.dart/builders/pub-win-be/builds/2695/steps/pub%20tests/logs/stdio
Any idea what's going on here?
| infrastructure | serve warns on assets paths test is failing on the windows bot but not locally i just checked in a new test for pub it passes everywhere including on my local windows machine but it s failing on the bot with nbsp results of running pub serve nbsp nbsp process exited with exit code nbsp nbsp standard output nbsp nbsp msg serving myapp on nbsp nbsp msg build completed successfully nbsp nbsp standard error nbsp nbsp fine pub nbsp nbsp io spawning cmd c ver nbsp nbsp io finished ver exit code stdout microsoft windows nothing output on stderr nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web foo nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web foo assets i didn t think a negative exit code was even possible as you can see no exception is thrown it looks like the process just dies any idea what s going on here | 1 |
225,889 | 17,928,905,050 | IssuesEvent | 2021-09-10 06:16:38 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | Fix intermittent debugger test failure in beta3 staging branch | Type/Task Area/Debugger Area/IntegrationTest Points/0.25 SwanLakeDump | **Description:**
The following debugger integration test case seems to be failing intermittently due to an timeout issue.

**Describe your task(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Fix intermittent debugger test failure in beta3 staging branch - **Description:**
The following debugger integration test case seems to be failing intermittently due to an timeout issue.

**Describe your task(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| non_infrastructure | fix intermittent debugger test failure in staging branch description the following debugger integration test case seems to be failing intermittently due to an timeout issue describe your task s related issues optional suggested labels optional suggested assignees optional | 0 |
781,644 | 27,445,126,303 | IssuesEvent | 2023-03-02 13:48:21 | Testaustime/testaustime-frontend | https://api.github.com/repos/Testaustime/testaustime-frontend | closed | If there is no data, the day chart is broken | bug priority-medium | The chart should show 0s for all days (a line at the bottom instead of middle)

| 1.0 | If there is no data, the day chart is broken - The chart should show 0s for all days (a line at the bottom instead of middle)

| non_infrastructure | if there is no data the day chart is broken the chart should show for all days a line at the bottom instead of middle | 0 |
48,731 | 13,184,725,713 | IssuesEvent | 2020-08-12 19:58:54 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py (Trac #110) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/110
, reported by icecube and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-09-06T17:23:26",
"description": "I see a different output running in icerec (trunk, RC 1-6-0)\n'make test -C TWRDistillery' --> Failed\n(Troy recommended 'ctest -VV -R TWRDistillery::TWRrecoTest' --> Failed with \n/Users/schlenst/prog/ice/icerec/trunk/src/TWRDistillery/private/twr-distillery/I3TWRReTrigger.cxx:127: FATAL: No DMADD file) \nand \n'python TWRDistillery/resources/scripts/TWRrecoTest.py' --> OK\n\nWhats wrong? Thanks, Stefan",
"reporter": "icecube",
"cc": "",
"resolution": "invalid",
"_ts": "1189099406000000",
"component": "cmake",
"summary": "make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py",
"priority": "normal",
"keywords": "running tests",
"time": "2007-09-06T15:05:10",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py (Trac #110) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/110
, reported by icecube and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-09-06T17:23:26",
"description": "I see a different output running in icerec (trunk, RC 1-6-0)\n'make test -C TWRDistillery' --> Failed\n(Troy recommended 'ctest -VV -R TWRDistillery::TWRrecoTest' --> Failed with \n/Users/schlenst/prog/ice/icerec/trunk/src/TWRDistillery/private/twr-distillery/I3TWRReTrigger.cxx:127: FATAL: No DMADD file) \nand \n'python TWRDistillery/resources/scripts/TWRrecoTest.py' --> OK\n\nWhats wrong? Thanks, Stefan",
"reporter": "icecube",
"cc": "",
"resolution": "invalid",
"_ts": "1189099406000000",
"component": "cmake",
"summary": "make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py",
"priority": "normal",
"keywords": "running tests",
"time": "2007-09-06T15:05:10",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| non_infrastructure | make test c twrdistillery and python twrdistillery resources scripts twrrecotest py trac migrated from reported by icecube and owned by troy json status closed changetime description i see a different output running in icerec trunk rc n make test c twrdistillery failed n troy recommended ctest vv r twrdistillery twrrecotest failed with n users schlenst prog ice icerec trunk src twrdistillery private twr distillery cxx fatal no dmadd file nand n python twrdistillery resources scripts twrrecotest py ok n nwhats wrong thanks stefan reporter icecube cc resolution invalid ts component cmake summary make test c twrdistillery and python twrdistillery resources scripts twrrecotest py priority normal keywords running tests time milestone owner troy type defect | 0 |
69,324 | 8,393,438,250 | IssuesEvent | 2018-10-09 20:33:36 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | icon on Brave Rewards Panel needs link or purpose | QA/Test-Plan-Specified QA/Yes design feature/rewards | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
NOTE THAT THIS IS THE REPOSITORY FOR THE UPCOMING VERSION OF BRAVE. SEE [browser-laptop](https://github.com/brave/browser-laptop)FOR THE CURRENT PRODUCTION VERSION OF BRAVE ON MACOS, WINDOWS AND LINUX.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
There is an icon of an exclamation point in a circle on the Brave Rewards Panel that doesn't seem to have a purpose or link (it's not mentioned in the spec but it is implemented). Need to determine what this icon should link to or what it's purpose should be (if it's not always visible, when does it display?).
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. If not enabled, enable Rewards.
2. Navigate to a site.
3. Click on the BAT logo.
## Actual result:
Next to 'Include in Brave Auto-Contribute' text note the icon. Clicking on it doesn't do anything.

## Expected result:
Clicking on the icon or hovering over it should provide some information. Unless it's not always supposed to display in which case this should be documented.
## Reproduces how often:
easy
## Brave version (chrome://version info)
Brave | 0.55.11 Chromium: 70.0.3538.35 (Official Build) beta(64-bit)
-- | --
Revision | 28dcb499844fa40c28d5f62e337876cb936f79f5-refs/branch-heads/3538@{#678}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds? yes
- Does it reproduce on browser-laptop? n/a
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
| 1.0 | icon on Brave Rewards Panel needs link or purpose - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
NOTE THAT THIS IS THE REPOSITORY FOR THE UPCOMING VERSION OF BRAVE. SEE [browser-laptop](https://github.com/brave/browser-laptop)FOR THE CURRENT PRODUCTION VERSION OF BRAVE ON MACOS, WINDOWS AND LINUX.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
There is an icon of an exclamation point in a circle on the Brave Rewards Panel that doesn't seem to have a purpose or link (it's not mentioned in the spec but it is implemented). Need to determine what this icon should link to or what it's purpose should be (if it's not always visible, when does it display?).
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. If not enabled, enable Rewards.
2. Navigate to a site.
3. Click on the BAT logo.
## Actual result:
Next to 'Include in Brave Auto-Contribute' text note the icon. Clicking on it doesn't do anything.

## Expected result:
Clicking on the icon or hovering over it should provide some information. Unless it's not always supposed to display in which case this should be documented.
## Reproduces how often:
easy
## Brave version (chrome://version info)
Brave | 0.55.11 Chromium: 70.0.3538.35 (Official Build) beta(64-bit)
-- | --
Revision | 28dcb499844fa40c28d5f62e337876cb936f79f5-refs/branch-heads/3538@{#678}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds? yes
- Does it reproduce on browser-laptop? n/a
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
| non_infrastructure | icon on brave rewards panel needs link or purpose have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue note that this is the repository for the upcoming version of brave see the current production version of brave on macos windows and linux please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description there is an icon of an exclamation point in a circle on the brave rewards panel that doesn t seem to have a purpose or link it s not mentioned in the spec but it is implemented need to determine what this icon should link to or what it s purpose should be if it s not always visible when does it display steps to reproduce if not enabled enable rewards navigate to a site click on the bat logo actual result next to include in brave auto contribute text note the icon clicking on it doesn t do anything expected result clicking on the icon or hovering over it should provide some information unless it s not always supposed to display in which case this should be documented reproduces how often easy brave version chrome version info brave chromium official build beta bit revision refs branch heads os mac os x reproducible on current release does it reproduce on brave browser dev beta builds yes does it reproduce on browser laptop n a website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information | 0 |
675,371 | 23,092,430,126 | IssuesEvent | 2022-07-26 16:15:13 | Kong/kubernetes-ingress-controller | https://api.github.com/repos/Kong/kubernetes-ingress-controller | closed | Regression Testing with Kuma | priority/medium area/ci | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Problem Statement
Changes to helm chart will occasionally break deployment of KIC when a service mesh is present. Having a regression tests greatly reduces risk of breaking changes.
### Proposed Solution
Create an automated deployment of GW, KIC and Kuma when a PR is created. If the environment stands up, proceed to next phase of the release cycle.
### Additional information
- [ ] needs https://github.com/Kong/kubernetes-testing-framework/issues/303
### Acceptance Criteria
As an end user, I consider it acceptable when you can deploy GW, KIC and Kuma without having to make unreasonable changes to the default values.yaml. | 1.0 | Regression Testing with Kuma - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Problem Statement
Changes to helm chart will occasionally break deployment of KIC when a service mesh is present. Having a regression tests greatly reduces risk of breaking changes.
### Proposed Solution
Create an automated deployment of GW, KIC and Kuma when a PR is created. If the environment stands up, proceed to next phase of the release cycle.
### Additional information
- [ ] needs https://github.com/Kong/kubernetes-testing-framework/issues/303
### Acceptance Criteria
As an end user, I consider it acceptable when you can deploy GW, KIC and Kuma without having to make unreasonable changes to the default values.yaml. | non_infrastructure | regression testing with kuma is there an existing issue for this i have searched the existing issues problem statement changes to helm chart will occasionally break deployment of kic when a service mesh is present having a regression tests greatly reduces risk of breaking changes proposed solution create an automated deployment of gw kic and kuma when a pr is created if the environment stands up proceed to next phase of the release cycle additional information needs acceptance criteria as an end user i consider it acceptable when you can deploy gw kic and kuma without having to make unreasonable changes to the default values yaml | 0 |
8,282 | 7,324,877,288 | IssuesEvent | 2018-03-03 01:41:44 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | [uap] CoreFx build is failing since 'RemotelyInvokable' does not contain a definition for 'LongWait' | area-Infrastructure | https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1436514
```text
2018-03-02T21:25:47.5255290Z Build FAILED.
2018-03-02T21:25:47.5268140Z
2018-03-02T21:25:47.5269301Z ProcessTestBase.Uap.cs(68,57): error CS0117: 'RemotelyInvokable' does not contain a definition for 'LongWait' [E:\A\_work\36\s\corefx\src\System.Diagnostics.Process\tests\System.Diagnostics.Process.Tests.csproj]
2018-03-02T21:25:47.5272048Z 0 Warning(s)
2018-03-02T21:25:47.5273015Z 1 Error(s)
```
cc @jkotas, @stephentoub | 1.0 | [uap] CoreFx build is failing since 'RemotelyInvokable' does not contain a definition for 'LongWait' - https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1436514
```text
2018-03-02T21:25:47.5255290Z Build FAILED.
2018-03-02T21:25:47.5268140Z
2018-03-02T21:25:47.5269301Z ProcessTestBase.Uap.cs(68,57): error CS0117: 'RemotelyInvokable' does not contain a definition for 'LongWait' [E:\A\_work\36\s\corefx\src\System.Diagnostics.Process\tests\System.Diagnostics.Process.Tests.csproj]
2018-03-02T21:25:47.5272048Z 0 Warning(s)
2018-03-02T21:25:47.5273015Z 1 Error(s)
```
cc @jkotas, @stephentoub | infrastructure | corefx build is failing since remotelyinvokable does not contain a definition for longwait text build failed processtestbase uap cs error remotelyinvokable does not contain a definition for longwait warning s error s cc jkotas stephentoub | 1 |
35,710 | 14,852,460,350 | IssuesEvent | 2021-01-18 08:37:31 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Need clarification whether the feature can be disabled/removed from a cluster | Pri2 container-service/svc cxp doc-enhancement triaged | There is no mention of whether the feature can be removed from a cluster and I saw that there's no ` --disable-azure-rbac` command in the CLI.
Can this be clarified in the docs please?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 42afd8de-b697-100c-98c7-ae8e6a6f1b92
* Version Independent ID: 81674054-0ea2-210b-2e65-18a0f15617c1
* Content: [Manage Azure RBAC in Kubernetes From Azure - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/manage-azure-rbac)
* Content Source: [articles/aks/manage-azure-rbac.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/manage-azure-rbac.md)
* Service: **container-service**
* GitHub Login: @palma21
* Microsoft Alias: **jpalma** | 1.0 | Need clarification whether the feature can be disabled/removed from a cluster - There is no mention of whether the feature can be removed from a cluster and I saw that there's no ` --disable-azure-rbac` command in the CLI.
Can this be clarified in the docs please?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 42afd8de-b697-100c-98c7-ae8e6a6f1b92
* Version Independent ID: 81674054-0ea2-210b-2e65-18a0f15617c1
* Content: [Manage Azure RBAC in Kubernetes From Azure - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/manage-azure-rbac)
* Content Source: [articles/aks/manage-azure-rbac.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/manage-azure-rbac.md)
* Service: **container-service**
* GitHub Login: @palma21
* Microsoft Alias: **jpalma** | non_infrastructure | need clarification whether the feature can be disabled removed from a cluster there is no mention of whether the feature can be removed from a cluster and i saw that there s no disable azure rbac command in the cli can this be clarified in the docs please document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login microsoft alias jpalma | 0 |
619,618 | 19,531,195,562 | IssuesEvent | 2021-12-30 17:09:04 | jshmrtn/hygeia | https://api.github.com/repos/jshmrtn/hygeia | closed | Befreiung von der Quarantäne Text anpassen | enhancement high-priority | ## Is your feature request related to a problem? Please describe.
In der Übersicht von einer KP steht dass man sich aus der Quarantäne entlassen kann, wenn man innerhalb von 6 Monaten bereits erkrankt ist. (siehe Screenshot)
## Describe the solution you'd like
Bitte den ersten Satz anpassen auf:
"Sie können sich von der Quarantäne befreien lassen, wenn Sie im Besitz eines gültigen Covid-Zertifikates sind."

## Describe alternatives you've considered
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
## Additional context
<!--
Add any other context or screenshots about the feature request here.
-->
## References
* https://app.forecast.it/project/P-205/scoping/T3283 | 1.0 | Befreiung von der Quarantäne Text anpassen - ## Is your feature request related to a problem? Please describe.
In der Übersicht von einer KP steht dass man sich aus der Quarantäne entlassen kann, wenn man innerhalb von 6 Monaten bereits erkrankt ist. (siehe Screenshot)
## Describe the solution you'd like
Bitte den ersten Satz anpassen auf:
"Sie können sich von der Quarantäne befreien lassen, wenn Sie im Besitz eines gültigen Covid-Zertifikates sind."

## Describe alternatives you've considered
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
## Additional context
<!--
Add any other context or screenshots about the feature request here.
-->
## References
* https://app.forecast.it/project/P-205/scoping/T3283 | non_infrastructure | befreiung von der quarantäne text anpassen is your feature request related to a problem please describe in der übersicht von einer kp steht dass man sich aus der quarantäne entlassen kann wenn man innerhalb von monaten bereits erkrankt ist siehe screenshot describe the solution you d like bitte den ersten satz anpassen auf sie können sich von der quarantäne befreien lassen wenn sie im besitz eines gültigen covid zertifikates sind describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here references | 0 |
37,619 | 10,041,088,038 | IssuesEvent | 2019-07-18 21:41:06 | gunrock/gunrock | https://api.github.com/repos/gunrock/gunrock | opened | Too many warnings during build | bug build issue enhancement | **Summary:**
We have way too many warnings building Gunrock. They've lost their utility.
**Reproducibility:**
Always
**To Reproduce:**
Build Gunrock as usual.
**Expected Behavior:**
There should be absolutely no warnings. | 1.0 | Too many warnings during build - **Summary:**
We have way too many warnings building Gunrock. They've lost their utility.
**Reproducibility:**
Always
**To Reproduce:**
Build Gunrock as usual.
**Expected Behavior:**
There should be absolutely no warnings. | non_infrastructure | too many warnings during build summary we have way too many warnings building gunrock they ve lost their utility reproducibility always to reproduce build gunrock as usual expected behavior there should be absolutely no warnings | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.